[GitHub] [spark] ulysses-you commented on a diff in pull request #40589: [SPARK-38697][SQL] Extend SparkSessionExtensions to inject rules into AQE query stage optimizer

2023-03-30 Thread via GitHub
ulysses-you commented on code in PR #40589: URL: https://github.com/apache/spark/pull/40589#discussion_r1154046952 ## sql/core/src/test/scala/org/apache/spark/sql/SparkSessionExtensionSuite.scala: ## @@ -1161,3 +1177,12 @@ object AddLimit extends Rule[LogicalPlan] { case _

[GitHub] [spark] HeartSaVioR commented on a diff in pull request #40561: [SPARK-42931][SS] Introduce dropDuplicatesWithinWatermark

2023-03-30 Thread via GitHub
HeartSaVioR commented on code in PR #40561: URL: https://github.com/apache/spark/pull/40561#discussion_r1154042557 ## sql/core/src/test/scala/org/apache/spark/sql/streaming/StreamingDeduplicationWithinWatermarkSuite.scala: ## @@ -0,0 +1,184 @@ +/* + * Licensed to the Apache

[GitHub] [spark] juanvisoler commented on a diff in pull request #40608: [SPARK-35198][CORE][PYTHON][SQL] Add support for calling debugCodegen from Python & Java

2023-03-30 Thread via GitHub
juanvisoler commented on code in PR #40608: URL: https://github.com/apache/spark/pull/40608#discussion_r1154034968 ## python/pyspark/sql/dataframe.py: ## @@ -706,6 +706,25 @@ def explain( assert self._sc._jvm is not None

[GitHub] [spark] ulysses-you commented on a diff in pull request #40589: [SPARK-38697][SQL] Extend SparkSessionExtensions to inject rules into AQE query stage optimizer

2023-03-30 Thread via GitHub
ulysses-you commented on code in PR #40589: URL: https://github.com/apache/spark/pull/40589#discussion_r1154034136 ## sql/core/src/main/scala/org/apache/spark/sql/SparkSessionExtensions.scala: ## @@ -111,11 +112,12 @@ class SparkSessionExtensions { type FunctionDescription =

[GitHub] [spark] cloud-fan commented on pull request #40258: [SPARK-42655][SQL] Incorrect ambiguous column reference error

2023-03-30 Thread via GitHub
cloud-fan commented on PR #40258: URL: https://github.com/apache/spark/pull/40258#issuecomment-1491283077 according to the [code in 2.3](https://github.com/apache/spark/blob/branch-2.3/sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/plans/logical/LogicalPlan.scala#L190), I think

[GitHub] [spark] cloud-fan commented on pull request #40258: [SPARK-42655][SQL] Incorrect ambiguous column reference error

2023-03-30 Thread via GitHub
cloud-fan commented on PR #40258: URL: https://github.com/apache/spark/pull/40258#issuecomment-1491280956 > FWIW Both the use cases were working fine in Spark 2.3 Sorry I missed this point. Do you know how it worked in 2.3? Did 2.3 also call `distinct` before returning the result?

[GitHub] [spark] ivoson commented on a diff in pull request #40610: [SPARK-42626][CONNECT] Add Destructive Iterator for SparkResult

2023-03-30 Thread via GitHub
ivoson commented on code in PR #40610: URL: https://github.com/apache/spark/pull/40610#discussion_r1154005920 ## connector/connect/client/jvm/src/main/scala/org/apache/spark/sql/connect/client/SparkResult.scala: ## @@ -134,24 +134,41 @@ private[sql] class SparkResult[T]( /**

[GitHub] [spark] yaooqinn commented on a diff in pull request #40583: [SPARK-42955][SQL] Skip classifyException and wrap AnalysisException for SparkThrowable

2023-03-30 Thread via GitHub
yaooqinn commented on code in PR #40583: URL: https://github.com/apache/spark/pull/40583#discussion_r1153995742 ## sql/catalyst/src/main/java/org/apache/spark/sql/connector/catalog/TableChange.java: ## @@ -628,6 +630,16 @@ public int hashCode() { result = 31 * result +

[GitHub] [spark] cloud-fan commented on a diff in pull request #40583: [SPARK-42955][SQL] Skip classifyException and wrap AnalysisException for SparkThrowable

2023-03-30 Thread via GitHub
cloud-fan commented on code in PR #40583: URL: https://github.com/apache/spark/pull/40583#discussion_r1153994592 ## sql/catalyst/src/main/java/org/apache/spark/sql/connector/catalog/TableChange.java: ## @@ -628,6 +630,16 @@ public int hashCode() { result = 31 * result +

[GitHub] [spark] cloud-fan closed pull request #40116: [SPARK-41391][SQL] The output column name of groupBy.agg(count_distinct) is incorrect

2023-03-30 Thread via GitHub
cloud-fan closed pull request #40116: [SPARK-41391][SQL] The output column name of groupBy.agg(count_distinct) is incorrect URL: https://github.com/apache/spark/pull/40116 -- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use

[GitHub] [spark] cloud-fan commented on pull request #40116: [SPARK-41391][SQL] The output column name of groupBy.agg(count_distinct) is incorrect

2023-03-30 Thread via GitHub
cloud-fan commented on PR #40116: URL: https://github.com/apache/spark/pull/40116#issuecomment-1491261819 thanks, merging to master! -- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific

[GitHub] [spark] dongjoon-hyun commented on a diff in pull request #40589: [SPARK-38697][SQL] Extend SparkSessionExtensions to inject rules into AQE query stage optimizer

2023-03-30 Thread via GitHub
dongjoon-hyun commented on code in PR #40589: URL: https://github.com/apache/spark/pull/40589#discussion_r1153990472 ## sql/core/src/main/scala/org/apache/spark/sql/SparkSessionExtensions.scala: ## @@ -111,11 +112,12 @@ class SparkSessionExtensions { type FunctionDescription

[GitHub] [spark] dongjoon-hyun commented on a diff in pull request #40589: [SPARK-38697][SQL] Extend SparkSessionExtensions to inject rules into AQE query stage optimizer

2023-03-30 Thread via GitHub
dongjoon-hyun commented on code in PR #40589: URL: https://github.com/apache/spark/pull/40589#discussion_r1153989800 ## sql/core/src/main/scala/org/apache/spark/sql/SparkSessionExtensions.scala: ## @@ -111,11 +112,12 @@ class SparkSessionExtensions { type FunctionDescription

[GitHub] [spark] dongjoon-hyun commented on a diff in pull request #40589: [SPARK-38697][SQL] Extend SparkSessionExtensions to inject rules into AQE query stage optimizer

2023-03-30 Thread via GitHub
dongjoon-hyun commented on code in PR #40589: URL: https://github.com/apache/spark/pull/40589#discussion_r1153988178 ## sql/core/src/main/scala/org/apache/spark/sql/SparkSessionExtensions.scala: ## @@ -111,11 +112,12 @@ class SparkSessionExtensions { type FunctionDescription

[GitHub] [spark] LuciferYang commented on pull request #40605: [SPARK-42958][CONNECT] Refactor `connect-jvm-client-mima-check` to support mima check with avro module

2023-03-30 Thread via GitHub
LuciferYang commented on PR #40605: URL: https://github.com/apache/spark/pull/40605#issuecomment-1491252507 GA passed -- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. To

[GitHub] [spark] LuciferYang commented on pull request #40610: [SPARK-42626][CONNECT] Add Destructive Iterator for SparkResult

2023-03-30 Thread via GitHub
LuciferYang commented on PR #40610: URL: https://github.com/apache/spark/pull/40610#issuecomment-1491250312 ``` 2023-03-30T16:09:39.936Z [info] - Dataset result destructive iterator *** FAILED *** (84 milliseconds) 2023-03-30T16:09:39.9382605Z

[GitHub] [spark] dongjoon-hyun commented on a diff in pull request #40589: [SPARK-38697][SQL] Extend SparkSessionExtensions to inject rules into AQE query stage optimizer

2023-03-30 Thread via GitHub
dongjoon-hyun commented on code in PR #40589: URL: https://github.com/apache/spark/pull/40589#discussion_r1153985168 ## sql/core/src/test/scala/org/apache/spark/sql/SparkSessionExtensionSuite.scala: ## @@ -1161,3 +1177,12 @@ object AddLimit extends Rule[LogicalPlan] { case

[GitHub] [spark] yaooqinn commented on pull request #40583: [SPARK-42955][SQL] Skip classifyException and wrap AnalysisException for SparkThrowable

2023-03-30 Thread via GitHub
yaooqinn commented on PR #40583: URL: https://github.com/apache/spark/pull/40583#issuecomment-1491249761 cc @cloud-fan @HyukjinKwon -- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific

[GitHub] [spark] dongjoon-hyun commented on a diff in pull request #40589: [SPARK-38697][SQL] Extend SparkSessionExtensions to inject rules into AQE query stage optimizer

2023-03-30 Thread via GitHub
dongjoon-hyun commented on code in PR #40589: URL: https://github.com/apache/spark/pull/40589#discussion_r1153984929 ## sql/core/src/test/scala/org/apache/spark/sql/SparkSessionExtensionSuite.scala: ## @@ -500,6 +500,22 @@ class SparkSessionExtensionSuite extends SparkFunSuite

[GitHub] [spark] hvanhovell commented on a diff in pull request #40611: [SPARK-42981][CONNECT] Add direct arrow serialization

2023-03-30 Thread via GitHub
hvanhovell commented on code in PR #40611: URL: https://github.com/apache/spark/pull/40611#discussion_r1153984573 ## connector/connect/client/jvm/src/main/scala/org/apache/spark/sql/connect/client/arrow/ArrowSerializer.scala: ## @@ -0,0 +1,529 @@ +/* + * Licensed to the Apache

[GitHub] [spark] cloud-fan commented on pull request #32987: [SPARK-35564][SQL] Support subexpression elimination for conditionally evaluated expressions

2023-03-30 Thread via GitHub
cloud-fan commented on PR #32987: URL: https://github.com/apache/spark/pull/32987#issuecomment-1491247967 After taking another thought, I think the idea is valid. If a subexpression will be evaluated at least once, and likely more than once due to conditional branches, it should be

[GitHub] [spark] cloud-fan commented on a diff in pull request #40602: [SPARK-42978][SQL] Derby: RENAME cannot qualify a new-table-Name with a schema-Name

2023-03-30 Thread via GitHub
cloud-fan commented on code in PR #40602: URL: https://github.com/apache/spark/pull/40602#discussion_r1153980662 ## sql/core/src/main/scala/org/apache/spark/sql/jdbc/DB2Dialect.scala: ## @@ -113,8 +114,9 @@ private object DB2Dialect extends JdbcDialect { // scalastyle:off

[GitHub] [spark] hvanhovell commented on a diff in pull request #40610: [SPARK-42626][CONNECT] Add Destructive Iterator for SparkResult

2023-03-30 Thread via GitHub
hvanhovell commented on code in PR #40610: URL: https://github.com/apache/spark/pull/40610#discussion_r1153980266 ## connector/connect/client/jvm/src/main/scala/org/apache/spark/sql/connect/client/SparkResult.scala: ## @@ -134,24 +134,41 @@ private[sql] class SparkResult[T](

[GitHub] [spark] cloud-fan commented on a diff in pull request #40602: [SPARK-42978][SQL] Derby: RENAME cannot qualify a new-table-Name with a schema-Name

2023-03-30 Thread via GitHub
cloud-fan commented on code in PR #40602: URL: https://github.com/apache/spark/pull/40602#discussion_r1153979914 ## sql/core/src/main/scala/org/apache/spark/sql/jdbc/JdbcDialects.scala: ## @@ -398,10 +398,24 @@ abstract class JdbcDialect extends Serializable with Logging {

[GitHub] [spark] HeartSaVioR commented on a diff in pull request #40561: [SPARK-42931][SS] Introduce dropDuplicatesWithinWatermark

2023-03-30 Thread via GitHub
HeartSaVioR commented on code in PR #40561: URL: https://github.com/apache/spark/pull/40561#discussion_r1153976547 ## sql/core/src/main/scala/org/apache/spark/sql/execution/streaming/statefulOperators.scala: ## @@ -980,3 +1022,65 @@ object StreamingDeduplicateExec { private

[GitHub] [spark] cloud-fan commented on a diff in pull request #40545: [SPARK-42918] Generalize handling of metadata attributes in FileSourceStrategy

2023-03-30 Thread via GitHub
cloud-fan commented on code in PR #40545: URL: https://github.com/apache/spark/pull/40545#discussion_r1153976307 ## sql/core/src/main/scala/org/apache/spark/sql/execution/datasources/FileSourceStrategy.scala: ## @@ -220,9 +220,20 @@ object FileSourceStrategy extends Strategy

[GitHub] [spark] Hisoka-X commented on a diff in pull request #40609: [SPARK-42316][SQL] Assign name to _LEGACY_ERROR_TEMP_2044

2023-03-30 Thread via GitHub
Hisoka-X commented on code in PR #40609: URL: https://github.com/apache/spark/pull/40609#discussion_r1153975175 ## sql/core/src/test/scala/org/apache/spark/sql/errors/QueryExecutionErrorsSuite.scala: ## @@ -625,6 +625,20 @@ class QueryExecutionErrorsSuite } } +

[GitHub] [spark] HeartSaVioR commented on a diff in pull request #40561: [SPARK-42931][SS] Introduce dropDuplicatesWithinWatermark

2023-03-30 Thread via GitHub
HeartSaVioR commented on code in PR #40561: URL: https://github.com/apache/spark/pull/40561#discussion_r1153973985 ## sql/core/src/test/scala/org/apache/spark/sql/streaming/StreamingDeduplicationWithinWatermarkSuite.scala: ## @@ -0,0 +1,184 @@ +/* + * Licensed to the Apache

[GitHub] [spark] Hisoka-X commented on a diff in pull request #40609: [SPARK-42316][SQL] Assign name to _LEGACY_ERROR_TEMP_2044

2023-03-30 Thread via GitHub
Hisoka-X commented on code in PR #40609: URL: https://github.com/apache/spark/pull/40609#discussion_r1153973500 ## sql/core/src/test/scala/org/apache/spark/sql/errors/QueryExecutionErrorsSuite.scala: ## @@ -625,6 +625,20 @@ class QueryExecutionErrorsSuite } } +

[GitHub] [spark] gengliangwang commented on pull request #40601: [SPARK-42975][SQL] Cast result type to timestamp type for string +/- interval

2023-03-30 Thread via GitHub
gengliangwang commented on PR #40601: URL: https://github.com/apache/spark/pull/40601#issuecomment-1491232046 > My suggestion is don't touch it to keep legacy workloads running. We should update the SQL queries to not use String so extensively. +1, totally agree! -- This is an

[GitHub] [spark] wangyum commented on pull request #40601: [SPARK-42975][SQL] Cast result type to timestamp type for string +/- interval

2023-03-30 Thread via GitHub
wangyum commented on PR #40601: URL: https://github.com/apache/spark/pull/40601#issuecomment-1491229091 +1 for fail it in ANSI mode. -- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific

[GitHub] [spark] gengliangwang commented on a diff in pull request #40561: [SPARK-42931][SS] Introduce dropDuplicatesWithinWatermark

2023-03-30 Thread via GitHub
gengliangwang commented on code in PR #40561: URL: https://github.com/apache/spark/pull/40561#discussion_r1153965249 ## sql/core/src/main/scala/org/apache/spark/sql/execution/streaming/statefulOperators.scala: ## @@ -980,3 +1022,65 @@ object StreamingDeduplicateExec {

[GitHub] [spark] gengliangwang commented on a diff in pull request #40561: [SPARK-42931][SS] Introduce dropDuplicatesWithinWatermark

2023-03-30 Thread via GitHub
gengliangwang commented on code in PR #40561: URL: https://github.com/apache/spark/pull/40561#discussion_r1153964890 ## sql/core/src/test/scala/org/apache/spark/sql/streaming/StreamingDeduplicationWithinWatermarkSuite.scala: ## @@ -0,0 +1,184 @@ +/* + * Licensed to the Apache

[GitHub] [spark] LuciferYang commented on a diff in pull request #36529: [SPARK-39102][CORE][SQL][DSTREAM] Add checkstyle rules to disabled use of Guava's `Files.createTempDir()`

2023-03-30 Thread via GitHub
LuciferYang commented on code in PR #36529: URL: https://github.com/apache/spark/pull/36529#discussion_r1153926757 ## common/network-common/src/main/java/org/apache/spark/network/util/JavaUtils.java: ## @@ -362,6 +364,60 @@ public static byte[] bufferToArray(ByteBuffer buffer)

[GitHub] [spark] LuciferYang commented on a diff in pull request #40613: [SPARK-42974][CORE] Restore `Utils.createTempDir` to use the `ShutdownHookManager` and clean up `JavaUtils.createTempDir` meth

2023-03-30 Thread via GitHub
LuciferYang commented on code in PR #40613: URL: https://github.com/apache/spark/pull/40613#discussion_r1153958746 ## streaming/src/test/java/test/org/apache/spark/streaming/JavaAPISuite.java: ## @@ -1476,7 +1476,7 @@ public void testCheckpointMasterRecovery() throws

[GitHub] [spark] LuciferYang commented on a diff in pull request #40613: [SPARK-42974][CORE] Restore `Utils.createTempDir` to use the `ShutdownHookManager` and clean up `JavaUtils.createTempDir` meth

2023-03-30 Thread via GitHub
LuciferYang commented on code in PR #40613: URL: https://github.com/apache/spark/pull/40613#discussion_r1153958604 ## core/src/test/java/test/org/apache/spark/JavaAPISuite.java: ## @@ -93,7 +94,7 @@ public class JavaAPISuite implements Serializable { @Before public void

[GitHub] [spark] LuciferYang commented on a diff in pull request #40613: [SPARK-42974][CORE] Restore `Utils.createTempDir` to use the `ShutdownHookManager` and clean up `JavaUtils.createTempDir` meth

2023-03-30 Thread via GitHub
LuciferYang commented on code in PR #40613: URL: https://github.com/apache/spark/pull/40613#discussion_r1153958391 ## core/src/test/java/test/org/apache/spark/Java8RDDAPISuite.java: ## @@ -246,7 +246,7 @@ public void mapPartitions() { @Test public void sequenceFile()

[GitHub] [spark] LuciferYang commented on a diff in pull request #40613: [SPARK-42974][CORE] Restore `Utils.createTempDir` to use the `ShutdownHookManager` and clean up `JavaUtils.createTempDir` meth

2023-03-30 Thread via GitHub
LuciferYang commented on code in PR #40613: URL: https://github.com/apache/spark/pull/40613#discussion_r1153957872 ## common/network-yarn/src/main/java/org/apache/spark/network/yarn/YarnShuffleService.java: ## @@ -243,7 +243,9 @@ protected void serviceInit(Configuration

[GitHub] [spark] LuciferYang commented on a diff in pull request #40613: [SPARK-42974][CORE] Restore `Utils.createTempDir` to use the `ShutdownHookManager` and clean up `JavaUtils.createTempDir` meth

2023-03-30 Thread via GitHub
LuciferYang commented on code in PR #40613: URL: https://github.com/apache/spark/pull/40613#discussion_r1153957701 ## common/network-shuffle/src/test/java/org/apache/spark/network/shuffle/TestShuffleDataContext.java: ## @@ -47,8 +47,9 @@ public TestShuffleDataContext(int

[GitHub] [spark] hvanhovell commented on a diff in pull request #40610: [SPARK-42626][CONNECT] Add Destructive Iterator for SparkResult

2023-03-30 Thread via GitHub
hvanhovell commented on code in PR #40610: URL: https://github.com/apache/spark/pull/40610#discussion_r1153957374 ## connector/connect/client/jvm/src/main/scala/org/apache/spark/sql/connect/client/SparkResult.scala: ## @@ -134,24 +134,41 @@ private[sql] class SparkResult[T](

[GitHub] [spark] hvanhovell commented on a diff in pull request #40610: [SPARK-42626][CONNECT] Add Destructive Iterator for SparkResult

2023-03-30 Thread via GitHub
hvanhovell commented on code in PR #40610: URL: https://github.com/apache/spark/pull/40610#discussion_r1153957184 ## connector/connect/client/jvm/src/main/scala/org/apache/spark/sql/connect/client/SparkResult.scala: ## @@ -45,7 +45,7 @@ private[sql] class SparkResult[T](

[GitHub] [spark] LuciferYang commented on a diff in pull request #40613: [SPARK-42974][CORE] Restore `Utils.createTempDir` to use the `ShutdownHookManager` and clean up `JavaUtils.createTempDir` meth

2023-03-30 Thread via GitHub
LuciferYang commented on code in PR #40613: URL: https://github.com/apache/spark/pull/40613#discussion_r1153956628 ## common/network-shuffle/src/test/java/org/apache/spark/network/shuffle/ExternalBlockHandlerSuite.java: ## @@ -125,7 +125,7 @@ private void checkDiagnosisResult(

[GitHub] [spark] LuciferYang commented on a diff in pull request #40613: [SPARK-42974][CORE] Restore `Utils.createTempDir` to use the `ShutdownHookManager` and clean up `JavaUtils.createTempDir` meth

2023-03-30 Thread via GitHub
LuciferYang commented on code in PR #40613: URL: https://github.com/apache/spark/pull/40613#discussion_r1153955764 ## common/network-common/src/test/java/org/apache/spark/network/StreamTestHelper.java: ## @@ -49,7 +49,7 @@ private static ByteBuffer createBuffer(int bufSize) {

[GitHub] [spark] HeartSaVioR commented on a diff in pull request #40561: [SPARK-42931][SS] Introduce dropDuplicatesWithinWatermark

2023-03-30 Thread via GitHub
HeartSaVioR commented on code in PR #40561: URL: https://github.com/apache/spark/pull/40561#discussion_r1153951716 ## sql/core/src/main/scala/org/apache/spark/sql/execution/streaming/statefulOperators.scala: ## @@ -980,3 +1022,65 @@ object StreamingDeduplicateExec { private

[GitHub] [spark] yaooqinn commented on pull request #40602: [SPARK-42978][SQL] Derby: RENAME cannot qualify a new-table-Name with a schema-Name

2023-03-30 Thread via GitHub
yaooqinn commented on PR #40602: URL: https://github.com/apache/spark/pull/40602#issuecomment-1491192780 cc @cloud-fan @HyukjinKwon thanks -- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the

[GitHub] [spark] LuciferYang commented on a diff in pull request #36529: [SPARK-39102][CORE][SQL][DSTREAM] Add checkstyle rules to disabled use of Guava's `Files.createTempDir()`

2023-03-30 Thread via GitHub
LuciferYang commented on code in PR #36529: URL: https://github.com/apache/spark/pull/36529#discussion_r1153926757 ## common/network-common/src/main/java/org/apache/spark/network/util/JavaUtils.java: ## @@ -362,6 +364,60 @@ public static byte[] bufferToArray(ByteBuffer buffer)

[GitHub] [spark] HeartSaVioR commented on a diff in pull request #40561: [SPARK-42931][SS] Introduce dropDuplicatesWithinWatermark

2023-03-30 Thread via GitHub
HeartSaVioR commented on code in PR #40561: URL: https://github.com/apache/spark/pull/40561#discussion_r1153936004 ## sql/core/src/main/scala/org/apache/spark/sql/execution/streaming/statefulOperators.scala: ## @@ -980,3 +1022,65 @@ object StreamingDeduplicateExec { private

[GitHub] [spark] HeartSaVioR commented on a diff in pull request #40561: [SPARK-42931][SS] Introduce dropDuplicatesWithinWatermark

2023-03-30 Thread via GitHub
HeartSaVioR commented on code in PR #40561: URL: https://github.com/apache/spark/pull/40561#discussion_r1153935792 ## sql/core/src/main/scala/org/apache/spark/sql/execution/streaming/statefulOperators.scala: ## @@ -980,3 +1022,65 @@ object StreamingDeduplicateExec { private

[GitHub] [spark] HeartSaVioR commented on a diff in pull request #40561: [SPARK-42931][SS] Introduce dropDuplicatesWithinWatermark

2023-03-30 Thread via GitHub
HeartSaVioR commented on code in PR #40561: URL: https://github.com/apache/spark/pull/40561#discussion_r1153935588 ## sql/core/src/main/scala/org/apache/spark/sql/execution/streaming/statefulOperators.scala: ## @@ -980,3 +1022,65 @@ object StreamingDeduplicateExec { private

[GitHub] [spark] LuciferYang commented on a diff in pull request #40613: [SPARK-42974][CORE] Separate Implementation of `Utils.createTempDir` and `JavaUtils.createTempDir`

2023-03-30 Thread via GitHub
LuciferYang commented on code in PR #40613: URL: https://github.com/apache/spark/pull/40613#discussion_r1153934642 ## core/src/main/scala/org/apache/spark/util/Utils.scala: ## @@ -320,7 +320,28 @@ private[spark] object Utils extends Logging { * newly created, and is not

[GitHub] [spark] LuciferYang closed pull request #40598: [SPARK-42974][CORE] Restore `Utils#createTempDir` use `ShutdownHookManager#registerShutdownDeleteDir` to cleanup tempDir

2023-03-30 Thread via GitHub
LuciferYang closed pull request #40598: [SPARK-42974][CORE] Restore `Utils#createTempDir` use `ShutdownHookManager#registerShutdownDeleteDir` to cleanup tempDir URL: https://github.com/apache/spark/pull/40598 -- This is an automated message from the Apache Git Service. To respond to the

[GitHub] [spark] HeartSaVioR commented on a diff in pull request #40561: [SPARK-42931][SS] Introduce dropDuplicatesWithinWatermark

2023-03-30 Thread via GitHub
HeartSaVioR commented on code in PR #40561: URL: https://github.com/apache/spark/pull/40561#discussion_r1153932172 ## sql/core/src/test/scala/org/apache/spark/sql/streaming/StreamingDeduplicationWithinWatermarkSuite.scala: ## @@ -0,0 +1,184 @@ +/* + * Licensed to the Apache

[GitHub] [spark] HeartSaVioR commented on a diff in pull request #40561: [SPARK-42931][SS] Introduce dropDuplicatesWithinWatermark

2023-03-30 Thread via GitHub
HeartSaVioR commented on code in PR #40561: URL: https://github.com/apache/spark/pull/40561#discussion_r1153932172 ## sql/core/src/test/scala/org/apache/spark/sql/streaming/StreamingDeduplicationWithinWatermarkSuite.scala: ## @@ -0,0 +1,184 @@ +/* + * Licensed to the Apache

[GitHub] [spark] RyanBerti opened a new pull request, #40615: [SPARK-16484][SQL] Add support for Datasketches HllSketch

2023-03-30 Thread via GitHub
RyanBerti opened a new pull request, #40615: URL: https://github.com/apache/spark/pull/40615 ### What changes were proposed in this pull request? This PR adds a new dependency on the datasketches-java project, and provides 3 new functions which utilize Datasketches HllSketch and Union

[GitHub] [spark] LuciferYang commented on a diff in pull request #36529: [SPARK-39102][CORE][SQL][DSTREAM] Add checkstyle rules to disabled use of Guava's `Files.createTempDir()`

2023-03-30 Thread via GitHub
LuciferYang commented on code in PR #36529: URL: https://github.com/apache/spark/pull/36529#discussion_r1153925550 ## common/network-common/src/main/java/org/apache/spark/network/util/JavaUtils.java: ## @@ -362,6 +364,60 @@ public static byte[] bufferToArray(ByteBuffer buffer)

[GitHub] [spark] HeartSaVioR commented on a diff in pull request #40561: [SPARK-42931][SS] Introduce dropDuplicatesWithinWatermark

2023-03-30 Thread via GitHub
HeartSaVioR commented on code in PR #40561: URL: https://github.com/apache/spark/pull/40561#discussion_r1153931344 ## sql/core/src/test/scala/org/apache/spark/sql/streaming/StreamingDeduplicationWithinWatermarkSuite.scala: ## @@ -0,0 +1,184 @@ +/* + * Licensed to the Apache

[GitHub] [spark] HeartSaVioR commented on a diff in pull request #40561: [SPARK-42931][SS] Introduce dropDuplicatesWithinWatermark

2023-03-30 Thread via GitHub
HeartSaVioR commented on code in PR #40561: URL: https://github.com/apache/spark/pull/40561#discussion_r1153930902 ## sql/core/src/test/scala/org/apache/spark/sql/streaming/StreamingDeduplicationWithinWatermarkSuite.scala: ## @@ -0,0 +1,184 @@ +/* + * Licensed to the Apache

[GitHub] [spark] LuciferYang commented on a diff in pull request #40613: Revert "[SPARK-39204][CORE] Change `Utils.createTempDir` and `Utils.createDirectory` call the same logic method in `JavaUtils`

2023-03-30 Thread via GitHub
LuciferYang commented on code in PR #40613: URL: https://github.com/apache/spark/pull/40613#discussion_r1153930483 ## core/src/main/scala/org/apache/spark/util/Utils.scala: ## @@ -320,7 +320,28 @@ private[spark] object Utils extends Logging { * newly created, and is not

[GitHub] [spark] lucaspompeun commented on pull request #40614: [SPARK-42987][DOCS] Correction of protobuf sql documentation

2023-03-30 Thread via GitHub
lucaspompeun commented on PR #40614: URL: https://github.com/apache/spark/pull/40614#issuecomment-1491167649 I'have corrected the problem that cause build error in github workflow -- This is an automated message from the Apache Git Service. To respond to the message, please log on to

[GitHub] [spark] cloud-fan commented on pull request #40601: [SPARK-42975][SQL] Cast result type to timestamp type for string +/- interval

2023-03-30 Thread via GitHub
cloud-fan commented on PR #40601: URL: https://github.com/apache/spark/pull/40601#issuecomment-1491166746 Or we should probably fail it in ANSI mode, cc @gengliangwang -- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use

[GitHub] [spark] cloud-fan commented on pull request #40601: [SPARK-42975][SQL] Cast result type to timestamp type for string +/- interval

2023-03-30 Thread via GitHub
cloud-fan commented on PR #40601: URL: https://github.com/apache/spark/pull/40601#issuecomment-1491166217 The change makes sense, but I'd say this is a legacy feature and the existing behavior doesn't make sense at all. For string +/- internal, the string can be timestamp, timestamp_ntz

[GitHub] [spark] LuciferYang commented on a diff in pull request #40613: Revert "[SPARK-39204][CORE] Change `Utils.createTempDir` and `Utils.createDirectory` call the same logic method in `JavaUtils`

2023-03-30 Thread via GitHub
LuciferYang commented on code in PR #40613: URL: https://github.com/apache/spark/pull/40613#discussion_r1153928232 ## core/src/main/scala/org/apache/spark/util/Utils.scala: ## @@ -330,7 +351,9 @@ private[spark] object Utils extends Logging { def createTempDir( root:

[GitHub] [spark] lucaspompeun opened a new pull request, #40614: correction of protobuf sql docuentation

2023-03-30 Thread via GitHub
lucaspompeun opened a new pull request, #40614: URL: https://github.com/apache/spark/pull/40614 ### What changes were proposed in this pull request? Correction of code highlights in SQL protobuf documentation. old version:

[GitHub] [spark] LuciferYang commented on a diff in pull request #36529: [SPARK-39102][CORE][SQL][DSTREAM] Add checkstyle rules to disabled use of Guava's `Files.createTempDir()`

2023-03-30 Thread via GitHub
LuciferYang commented on code in PR #36529: URL: https://github.com/apache/spark/pull/36529#discussion_r1153926757 ## common/network-common/src/main/java/org/apache/spark/network/util/JavaUtils.java: ## @@ -362,6 +364,60 @@ public static byte[] bufferToArray(ByteBuffer buffer)

[GitHub] [spark] LuciferYang opened a new pull request, #40613: Revert "[SPARK-39204][CORE] Change `Utils.createTempDir` and `Utils.createDirectory` call the same logic method in `JavaUtils`

2023-03-30 Thread via GitHub
LuciferYang opened a new pull request, #40613: URL: https://github.com/apache/spark/pull/40613 This reverts commit 5cb5d1fa66ad9d6e94beb17d3fda3a8f220bc371. ### What changes were proposed in this pull request? ### Why are the changes needed? ###

[GitHub] [spark] LuciferYang commented on a diff in pull request #36529: [SPARK-39102][CORE][SQL][DSTREAM] Add checkstyle rules to disabled use of Guava's `Files.createTempDir()`

2023-03-30 Thread via GitHub
LuciferYang commented on code in PR #36529: URL: https://github.com/apache/spark/pull/36529#discussion_r1153925550 ## common/network-common/src/main/java/org/apache/spark/network/util/JavaUtils.java: ## @@ -362,6 +364,60 @@ public static byte[] bufferToArray(ByteBuffer buffer)

[GitHub] [spark] LuciferYang commented on a diff in pull request #36529: [SPARK-39102][CORE][SQL][DSTREAM] Add checkstyle rules to disabled use of Guava's `Files.createTempDir()`

2023-03-30 Thread via GitHub
LuciferYang commented on code in PR #36529: URL: https://github.com/apache/spark/pull/36529#discussion_r1153920360 ## common/network-common/src/main/java/org/apache/spark/network/util/JavaUtils.java: ## @@ -362,6 +364,60 @@ public static byte[] bufferToArray(ByteBuffer buffer)

[GitHub] [spark] LuciferYang commented on a diff in pull request #36529: [SPARK-39102][CORE][SQL][DSTREAM] Add checkstyle rules to disabled use of Guava's `Files.createTempDir()`

2023-03-30 Thread via GitHub
LuciferYang commented on code in PR #36529: URL: https://github.com/apache/spark/pull/36529#discussion_r1153917326 ## common/network-common/src/main/java/org/apache/spark/network/util/JavaUtils.java: ## @@ -362,6 +364,60 @@ public static byte[] bufferToArray(ByteBuffer buffer)

[GitHub] [spark] srowen commented on a diff in pull request #36529: [SPARK-39102][CORE][SQL][DSTREAM] Add checkstyle rules to disabled use of Guava's `Files.createTempDir()`

2023-03-30 Thread via GitHub
srowen commented on code in PR #36529: URL: https://github.com/apache/spark/pull/36529#discussion_r1153918827 ## common/network-common/src/main/java/org/apache/spark/network/util/JavaUtils.java: ## @@ -362,6 +364,60 @@ public static byte[] bufferToArray(ByteBuffer buffer) {

[GitHub] [spark] srowen commented on a diff in pull request #36529: [SPARK-39102][CORE][SQL][DSTREAM] Add checkstyle rules to disabled use of Guava's `Files.createTempDir()`

2023-03-30 Thread via GitHub
srowen commented on code in PR #36529: URL: https://github.com/apache/spark/pull/36529#discussion_r1153918687 ## common/network-common/src/main/java/org/apache/spark/network/util/JavaUtils.java: ## @@ -362,6 +364,60 @@ public static byte[] bufferToArray(ByteBuffer buffer) {

[GitHub] [spark] sadikovi commented on a diff in pull request #36529: [SPARK-39102][CORE][SQL][DSTREAM] Add checkstyle rules to disabled use of Guava's `Files.createTempDir()`

2023-03-30 Thread via GitHub
sadikovi commented on code in PR #36529: URL: https://github.com/apache/spark/pull/36529#discussion_r1153918589 ## common/network-common/src/main/java/org/apache/spark/network/util/JavaUtils.java: ## @@ -362,6 +364,60 @@ public static byte[] bufferToArray(ByteBuffer buffer) {

[GitHub] [spark] LuciferYang commented on a diff in pull request #36529: [SPARK-39102][CORE][SQL][DSTREAM] Add checkstyle rules to disabled use of Guava's `Files.createTempDir()`

2023-03-30 Thread via GitHub
LuciferYang commented on code in PR #36529: URL: https://github.com/apache/spark/pull/36529#discussion_r1153917326 ## common/network-common/src/main/java/org/apache/spark/network/util/JavaUtils.java: ## @@ -362,6 +364,60 @@ public static byte[] bufferToArray(ByteBuffer buffer)

[GitHub] [spark] rangadi commented on a diff in pull request #40561: [SPARK-42931][SS] Introduce dropDuplicatesWithinWatermark

2023-03-30 Thread via GitHub
rangadi commented on code in PR #40561: URL: https://github.com/apache/spark/pull/40561#discussion_r1153915858 ## sql/core/src/test/scala/org/apache/spark/sql/streaming/StreamingDeduplicationWithinWatermarkSuite.scala: ## @@ -0,0 +1,184 @@ +/* + * Licensed to the Apache

[GitHub] [spark] rangadi commented on a diff in pull request #40561: [SPARK-42931][SS] Introduce dropDuplicatesWithinWatermark

2023-03-30 Thread via GitHub
rangadi commented on code in PR #40561: URL: https://github.com/apache/spark/pull/40561#discussion_r1153915713 ## sql/core/src/test/scala/org/apache/spark/sql/streaming/StreamingDeduplicationWithinWatermarkSuite.scala: ## @@ -0,0 +1,184 @@ +/* + * Licensed to the Apache

[GitHub] [spark] rangadi commented on a diff in pull request #40561: [SPARK-42931][SS] Introduce dropDuplicatesWithinWatermark

2023-03-30 Thread via GitHub
rangadi commented on code in PR #40561: URL: https://github.com/apache/spark/pull/40561#discussion_r1153915130 ## sql/core/src/test/scala/org/apache/spark/sql/streaming/StreamingDeduplicationWithinWatermarkSuite.scala: ## @@ -0,0 +1,184 @@ +/* + * Licensed to the Apache

[GitHub] [spark] rangadi commented on a diff in pull request #40561: [SPARK-42931][SS] Introduce dropDuplicatesWithinWatermark

2023-03-30 Thread via GitHub
rangadi commented on code in PR #40561: URL: https://github.com/apache/spark/pull/40561#discussion_r1153914621 ## sql/core/src/test/scala/org/apache/spark/sql/streaming/StreamingDeduplicationWithinWatermarkSuite.scala: ## @@ -0,0 +1,184 @@ +/* + * Licensed to the Apache

[GitHub] [spark] rangadi commented on a diff in pull request #40586: [SPARK-42939][SS][CONNECT] Core streaming Python API for Spark Connect

2023-03-30 Thread via GitHub
rangadi commented on code in PR #40586: URL: https://github.com/apache/spark/pull/40586#discussion_r1153908962 ## python/pyspark/sql/connect/session.py: ## @@ -14,6 +14,7 @@ # See the License for the specific language governing permissions and # limitations under the License.

[GitHub] [spark] rangadi commented on a diff in pull request #40586: [SPARK-42939][SS][CONNECT] Core streaming Python API for Spark Connect

2023-03-30 Thread via GitHub
rangadi commented on code in PR #40586: URL: https://github.com/apache/spark/pull/40586#discussion_r1153908437 ## python/pyspark/sql/connect/session.py: ## @@ -489,10 +495,6 @@ def sparkContext(self) -> Any: def streams(self) -> Any: raise

[GitHub] [spark] HeartSaVioR commented on a diff in pull request #40561: [SPARK-42931][SS] Introduce dropDuplicatesWithinWatermark

2023-03-30 Thread via GitHub
HeartSaVioR commented on code in PR #40561: URL: https://github.com/apache/spark/pull/40561#discussion_r1153906537 ## sql/core/src/test/scala/org/apache/spark/sql/streaming/StreamingDeduplicationWithinWatermarkSuite.scala: ## @@ -0,0 +1,184 @@ +/* + * Licensed to the Apache

[GitHub] [spark] HeartSaVioR commented on a diff in pull request #40561: [SPARK-42931][SS] Introduce dropDuplicatesWithinWatermark

2023-03-30 Thread via GitHub
HeartSaVioR commented on code in PR #40561: URL: https://github.com/apache/spark/pull/40561#discussion_r1153906537 ## sql/core/src/test/scala/org/apache/spark/sql/streaming/StreamingDeduplicationWithinWatermarkSuite.scala: ## @@ -0,0 +1,184 @@ +/* + * Licensed to the Apache

[GitHub] [spark] rangadi commented on a diff in pull request #40586: [SPARK-42939][SS][CONNECT] Core streaming Python API for Spark Connect

2023-03-30 Thread via GitHub
rangadi commented on code in PR #40586: URL: https://github.com/apache/spark/pull/40586#discussion_r1153904387 ## python/pyspark/sql/connect/readwriter.py: ## @@ -37,7 +37,7 @@ from pyspark.sql.connect._typing import ColumnOrName, OptionalPrimitiveType from

[GitHub] [spark] HeartSaVioR commented on a diff in pull request #40561: [SPARK-42931][SS] Introduce dropDuplicatesWithinWatermark

2023-03-30 Thread via GitHub
HeartSaVioR commented on code in PR #40561: URL: https://github.com/apache/spark/pull/40561#discussion_r1153895088 ## sql/core/src/test/scala/org/apache/spark/sql/streaming/StreamingDeduplicationWithinWatermarkSuite.scala: ## @@ -0,0 +1,184 @@ +/* + * Licensed to the Apache

[GitHub] [spark] HeartSaVioR commented on a diff in pull request #40561: [SPARK-42931][SS] Introduce dropDuplicatesWithinWatermark

2023-03-30 Thread via GitHub
HeartSaVioR commented on code in PR #40561: URL: https://github.com/apache/spark/pull/40561#discussion_r1153904044 ## sql/core/src/test/scala/org/apache/spark/sql/streaming/StreamingDeduplicationWithinWatermarkSuite.scala: ## @@ -0,0 +1,184 @@ +/* + * Licensed to the Apache

[GitHub] [spark] github-actions[bot] closed pull request #38732: [SPARK-41210][K8S] Window based executor failure tracking mechanism

2023-03-30 Thread via GitHub
github-actions[bot] closed pull request #38732: [SPARK-41210][K8S] Window based executor failure tracking mechanism URL: https://github.com/apache/spark/pull/38732 -- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL

[GitHub] [spark] github-actions[bot] commented on pull request #39130: [SPARK-xxxxx][DOCUMENTATION][PYTHON] Fix grammar in docstring for toDF().

2023-03-30 Thread via GitHub
github-actions[bot] commented on PR #39130: URL: https://github.com/apache/spark/pull/39130#issuecomment-1491126263 We're closing this PR because it hasn't been updated in a while. This isn't a judgement on the merit of the PR in any way. It's just a way of keeping the PR queue manageable.

[GitHub] [spark] github-actions[bot] closed pull request #39102: [SPARK-41555][SQL] Multi sparkSession should share single SQLAppStatusStore

2023-03-30 Thread via GitHub
github-actions[bot] closed pull request #39102: [SPARK-41555][SQL] Multi sparkSession should share single SQLAppStatusStore URL: https://github.com/apache/spark/pull/39102 -- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use

[GitHub] [spark] LuciferYang commented on a diff in pull request #36529: [SPARK-39102][CORE][SQL][DSTREAM] Add checkstyle rules to disabled use of Guava's `Files.createTempDir()`

2023-03-30 Thread via GitHub
LuciferYang commented on code in PR #36529: URL: https://github.com/apache/spark/pull/36529#discussion_r115394 ## common/network-common/src/main/java/org/apache/spark/network/util/JavaUtils.java: ## @@ -362,6 +364,60 @@ public static byte[] bufferToArray(ByteBuffer buffer)

[GitHub] [spark] WweiL commented on a diff in pull request #40586: [SPARK-42939][SS][CONNECT] Core streaming Python API for Spark Connect

2023-03-30 Thread via GitHub
WweiL commented on code in PR #40586: URL: https://github.com/apache/spark/pull/40586#discussion_r1153899173 ## connector/connect/server/src/main/scala/org/apache/spark/sql/connect/planner/SparkConnectPlanner.scala: ## @@ -2120,7 +2130,6 @@ class SparkConnectPlanner(val

[GitHub] [spark] zhengruifeng commented on pull request #40612: [SPARK-42969][CONNECT][TESTS] Fix the comparison the result with Arrow optimization enabled/disabled

2023-03-30 Thread via GitHub
zhengruifeng commented on PR #40612: URL: https://github.com/apache/spark/pull/40612#issuecomment-1491119721 LGTM -- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. To

[GitHub] [spark] HeartSaVioR commented on a diff in pull request #40561: [SPARK-42931][SS] Introduce dropDuplicatesWithinWatermark

2023-03-30 Thread via GitHub
HeartSaVioR commented on code in PR #40561: URL: https://github.com/apache/spark/pull/40561#discussion_r1153895796 ## sql/core/src/test/scala/org/apache/spark/sql/streaming/StreamingDeduplicationWithinWatermarkSuite.scala: ## @@ -0,0 +1,184 @@ +/* + * Licensed to the Apache

[GitHub] [spark] HeartSaVioR commented on a diff in pull request #40561: [SPARK-42931][SS] Introduce dropDuplicatesWithinWatermark

2023-03-30 Thread via GitHub
HeartSaVioR commented on code in PR #40561: URL: https://github.com/apache/spark/pull/40561#discussion_r1153897632 ## sql/core/src/test/scala/org/apache/spark/sql/streaming/StreamingDeduplicationWithinWatermarkSuite.scala: ## @@ -0,0 +1,184 @@ +/* + * Licensed to the Apache

[GitHub] [spark] HeartSaVioR commented on a diff in pull request #40561: [SPARK-42931][SS] Introduce dropDuplicatesWithinWatermark

2023-03-30 Thread via GitHub
HeartSaVioR commented on code in PR #40561: URL: https://github.com/apache/spark/pull/40561#discussion_r1153896951 ## sql/core/src/test/scala/org/apache/spark/sql/streaming/StreamingDeduplicationWithinWatermarkSuite.scala: ## @@ -0,0 +1,184 @@ +/* + * Licensed to the Apache

[GitHub] [spark] HeartSaVioR commented on a diff in pull request #40561: [SPARK-42931][SS] Introduce dropDuplicatesWithinWatermark

2023-03-30 Thread via GitHub
HeartSaVioR commented on code in PR #40561: URL: https://github.com/apache/spark/pull/40561#discussion_r1153896257 ## sql/core/src/test/scala/org/apache/spark/sql/streaming/StreamingDeduplicationWithinWatermarkSuite.scala: ## @@ -0,0 +1,184 @@ +/* + * Licensed to the Apache

[GitHub] [spark] HeartSaVioR commented on a diff in pull request #40561: [SPARK-42931][SS] Introduce dropDuplicatesWithinWatermark

2023-03-30 Thread via GitHub
HeartSaVioR commented on code in PR #40561: URL: https://github.com/apache/spark/pull/40561#discussion_r1153896106 ## sql/core/src/test/scala/org/apache/spark/sql/streaming/StreamingDeduplicationWithinWatermarkSuite.scala: ## @@ -0,0 +1,184 @@ +/* + * Licensed to the Apache

[GitHub] [spark] HeartSaVioR commented on a diff in pull request #40561: [SPARK-42931][SS] Introduce dropDuplicatesWithinWatermark

2023-03-30 Thread via GitHub
HeartSaVioR commented on code in PR #40561: URL: https://github.com/apache/spark/pull/40561#discussion_r1153895796 ## sql/core/src/test/scala/org/apache/spark/sql/streaming/StreamingDeduplicationWithinWatermarkSuite.scala: ## @@ -0,0 +1,184 @@ +/* + * Licensed to the Apache

[GitHub] [spark] itholic commented on pull request #39937: [SPARK-42309][SQL] Introduce `INCOMPATIBLE_DATA_TO_TABLE` and sub classes.

2023-03-30 Thread via GitHub
itholic commented on PR #39937: URL: https://github.com/apache/spark/pull/39937#issuecomment-1491110788 Test passed. @MaxGekk could you take a look when you find some time? -- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and

[GitHub] [spark] HeartSaVioR commented on a diff in pull request #40561: [SPARK-42931][SS] Introduce dropDuplicatesWithinWatermark

2023-03-30 Thread via GitHub
HeartSaVioR commented on code in PR #40561: URL: https://github.com/apache/spark/pull/40561#discussion_r1153895088 ## sql/core/src/test/scala/org/apache/spark/sql/streaming/StreamingDeduplicationWithinWatermarkSuite.scala: ## @@ -0,0 +1,184 @@ +/* + * Licensed to the Apache

[GitHub] [spark] HeartSaVioR commented on a diff in pull request #40561: [SPARK-42931][SS] Introduce dropDuplicatesWithinWatermark

2023-03-30 Thread via GitHub
HeartSaVioR commented on code in PR #40561: URL: https://github.com/apache/spark/pull/40561#discussion_r1153895088 ## sql/core/src/test/scala/org/apache/spark/sql/streaming/StreamingDeduplicationWithinWatermarkSuite.scala: ## @@ -0,0 +1,184 @@ +/* + * Licensed to the Apache

[GitHub] [spark] HyukjinKwon commented on a diff in pull request #40591: [SPARK-42950][CORE] Add exit code in SparkListenerApplicationEnd

2023-03-30 Thread via GitHub
HyukjinKwon commented on code in PR #40591: URL: https://github.com/apache/spark/pull/40591#discussion_r1153892668 ## core/src/main/scala/org/apache/spark/scheduler/SparkListener.scala: ## @@ -289,7 +289,8 @@ case class SparkListenerApplicationStart( driverAttributes:

[GitHub] [spark] HeartSaVioR commented on pull request #40561: [SPARK-42931][SS] Introduce dropDuplicatesWithinWatermark

2023-03-30 Thread via GitHub
HeartSaVioR commented on PR #40561: URL: https://github.com/apache/spark/pull/40561#issuecomment-1491105348 > What is the decision about batch support? I just added support of batch in the latest commit. It needs be more test coverage for batch query support so that's why we have new

  1   2   3   >