pan3793 commented on code in PR #36496:
URL: https://github.com/apache/spark/pull/36496#discussion_r874423234
##
sql/core/src/test/scala/org/apache/spark/sql/execution/columnar/InMemoryColumnarQuerySuite.scala:
##
@@ -563,4 +564,51 @@ class InMemoryColumnarQuerySuite extends
sadikovi commented on code in PR #36562:
URL: https://github.com/apache/spark/pull/36562#discussion_r874433212
##
sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/json/JsonInferSchema.scala:
##
@@ -30,29 +30,16 @@ import
Yikun commented on code in PR #36509:
URL: https://github.com/apache/spark/pull/36509#discussion_r874452951
##
python/pyspark/pandas/supported_api_gen.py:
##
@@ -0,0 +1,377 @@
+#
+# Licensed to the Apache Software Foundation (ASF) under one or more
+# contributor license
Yikun commented on code in PR #36509:
URL: https://github.com/apache/spark/pull/36509#discussion_r874452951
##
python/pyspark/pandas/supported_api_gen.py:
##
@@ -0,0 +1,377 @@
+#
+# Licensed to the Apache Software Foundation (ASF) under one or more
+# contributor license
gengliangwang commented on code in PR #36562:
URL: https://github.com/apache/spark/pull/36562#discussion_r874425426
##
sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/json/JsonInferSchema.scala:
##
@@ -30,29 +30,16 @@ import
Yikun commented on code in PR #36509:
URL: https://github.com/apache/spark/pull/36509#discussion_r874452951
##
python/pyspark/pandas/supported_api_gen.py:
##
@@ -0,0 +1,377 @@
+#
+# Licensed to the Apache Software Foundation (ASF) under one or more
+# contributor license
cloud-fan commented on code in PR #36562:
URL: https://github.com/apache/spark/pull/36562#discussion_r874403362
##
sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/csv/CSVInferSchema.scala:
##
@@ -178,7 +164,8 @@ class CSVInferSchema(val options: CSVOptions) extends
gengliangwang commented on code in PR #36562:
URL: https://github.com/apache/spark/pull/36562#discussion_r874474944
##
sql/catalyst/src/test/scala/org/apache/spark/sql/catalyst/util/TimestampFormatterSuite.scala:
##
@@ -456,4 +456,19 @@ class TimestampFormatterSuite extends
sadikovi commented on code in PR #36562:
URL: https://github.com/apache/spark/pull/36562#discussion_r874424921
##
sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/csv/CSVInferSchema.scala:
##
@@ -178,7 +164,8 @@ class CSVInferSchema(val options: CSVOptions) extends
LuciferYang commented on code in PR #36496:
URL: https://github.com/apache/spark/pull/36496#discussion_r874402726
##
sql/core/src/test/scala/org/apache/spark/sql/execution/columnar/InMemoryColumnarQuerySuite.scala:
##
@@ -563,4 +564,51 @@ class InMemoryColumnarQuerySuite
Yikun commented on code in PR #36509:
URL: https://github.com/apache/spark/pull/36509#discussion_r874452951
##
python/pyspark/pandas/supported_api_gen.py:
##
@@ -0,0 +1,377 @@
+#
+# Licensed to the Apache Software Foundation (ASF) under one or more
+# contributor license
gengliangwang commented on code in PR #36562:
URL: https://github.com/apache/spark/pull/36562#discussion_r874415392
##
sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/csv/CSVInferSchema.scala:
##
@@ -178,7 +164,8 @@ class CSVInferSchema(val options: CSVOptions)
Yikun commented on code in PR #36509:
URL: https://github.com/apache/spark/pull/36509#discussion_r874457721
##
python/pyspark/pandas/supported_api_gen.py:
##
@@ -0,0 +1,377 @@
+#
+# Licensed to the Apache Software Foundation (ASF) under one or more
+# contributor license
beobest2 commented on code in PR #36509:
URL: https://github.com/apache/spark/pull/36509#discussion_r874470001
##
python/pyspark/pandas/supported_api_gen.py:
##
@@ -0,0 +1,377 @@
+#
+# Licensed to the Apache Software Foundation (ASF) under one or more
+# contributor license
cloud-fan commented on code in PR #36295:
URL: https://github.com/apache/spark/pull/36295#discussion_r874584082
##
sql/catalyst/src/main/java/org/apache/spark/sql/connector/read/SupportsPushDownLimit.java:
##
@@ -21,8 +21,8 @@
/**
* A mix-in interface for {@link
HyukjinKwon commented on code in PR #36576:
URL: https://github.com/apache/spark/pull/36576#discussion_r874632251
##
sql/core/src/test/scala/org/apache/spark/sql/BloomFilterAggregateQuerySuite.scala:
##
@@ -35,23 +34,26 @@ class BloomFilterAggregateQuerySuite extends QueryTest
jackylee-ch commented on PR #36578:
URL: https://github.com/apache/spark/pull/36578#issuecomment-1128721543
Great job.
BTW, is it posibble for user to define the Description? Sometimes, SQLText
is actually too big to show in Description, then use defined text is very
helpful to find the
AmplabJenkins commented on PR #36561:
URL: https://github.com/apache/spark/pull/36561#issuecomment-1128547684
Can one of the admins verify this patch?
--
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to
panbingkun commented on PR #36540:
URL: https://github.com/apache/spark/pull/36540#issuecomment-1128547367
@MaxGekk ping
--
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
To
cloud-fan commented on code in PR #36295:
URL: https://github.com/apache/spark/pull/36295#discussion_r874589600
##
sql/core/src/main/scala/org/apache/spark/sql/execution/datasources/v2/V2ScanRelationPushDown.scala:
##
@@ -419,6 +420,72 @@ object V2ScanRelationPushDown extends
HyukjinKwon opened a new pull request, #36576:
URL: https://github.com/apache/spark/pull/36576
### What changes were proposed in this pull request?
This PR proposes:
1. Use the function registry in the Spark Session being used
2. Move function registration into `beforeAll`
HyukjinKwon commented on code in PR #36576:
URL: https://github.com/apache/spark/pull/36576#discussion_r874625022
##
sql/core/src/test/scala/org/apache/spark/sql/SQLQuerySuite.scala:
##
@@ -147,6 +147,9 @@ class SQLQuerySuite extends QueryTest with
SharedSparkSession with
HyukjinKwon commented on code in PR #36576:
URL: https://github.com/apache/spark/pull/36576#discussion_r874632251
##
sql/core/src/test/scala/org/apache/spark/sql/BloomFilterAggregateQuerySuite.scala:
##
@@ -35,23 +34,26 @@ class BloomFilterAggregateQuerySuite extends QueryTest
LuciferYang commented on code in PR #36578:
URL: https://github.com/apache/spark/pull/36578#discussion_r874671414
##
sql/core/src/main/scala/org/apache/spark/sql/execution/SQLExecution.scala:
##
@@ -82,7 +82,7 @@ object SQLExecution {
val redactedStr = Utils
LuciferYang commented on code in PR #36578:
URL: https://github.com/apache/spark/pull/36578#discussion_r874671414
##
sql/core/src/main/scala/org/apache/spark/sql/execution/SQLExecution.scala:
##
@@ -82,7 +82,7 @@ object SQLExecution {
val redactedStr = Utils
MaxGekk opened a new pull request, #36579:
URL: https://github.com/apache/spark/pull/36579
### What changes were proposed in this pull request?
Wrap values of SQL configs and datasource options in error messages by
double quotes. Added the `toDSOption()` method to `QueryErrorsBase` to
beliefer commented on code in PR #36295:
URL: https://github.com/apache/spark/pull/36295#discussion_r874727085
##
sql/core/src/main/scala/org/apache/spark/sql/execution/datasources/jdbc/JDBCRDD.scala:
##
@@ -304,10 +307,11 @@ private[jdbc] class JDBCRDD(
}
val
cloud-fan commented on PR #36572:
URL: https://github.com/apache/spark/pull/36572#issuecomment-1128539432
thanks for the review, merging to master/3.3!
--
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to
cloud-fan commented on code in PR #36531:
URL: https://github.com/apache/spark/pull/36531#discussion_r874528129
##
sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/expressions/Cast.scala:
##
@@ -2138,199 +2287,28 @@ case class Cast(
final override def
cloud-fan commented on code in PR #36295:
URL: https://github.com/apache/spark/pull/36295#discussion_r874586232
##
sql/core/src/main/scala/org/apache/spark/sql/execution/datasources/jdbc/JDBCRDD.scala:
##
@@ -304,10 +307,11 @@ private[jdbc] class JDBCRDD(
}
val
beliefer commented on code in PR #36295:
URL: https://github.com/apache/spark/pull/36295#discussion_r874600195
##
sql/catalyst/src/main/java/org/apache/spark/sql/connector/read/SupportsPushDownLimit.java:
##
@@ -21,8 +21,8 @@
/**
* A mix-in interface for {@link
gengliangwang commented on code in PR #36562:
URL: https://github.com/apache/spark/pull/36562#discussion_r874492296
##
sql/catalyst/src/test/scala/org/apache/spark/sql/catalyst/util/TimestampFormatterSuite.scala:
##
@@ -456,4 +456,19 @@ class TimestampFormatterSuite extends
HyukjinKwon commented on code in PR #36576:
URL: https://github.com/apache/spark/pull/36576#discussion_r874632251
##
sql/core/src/test/scala/org/apache/spark/sql/BloomFilterAggregateQuerySuite.scala:
##
@@ -35,23 +34,26 @@ class BloomFilterAggregateQuerySuite extends QueryTest
zero323 commented on PR #36547:
URL: https://github.com/apache/spark/pull/36547#issuecomment-1128565616
By itself LGTM. To concur with others, I also don't see shading issue and if
it there was one, we're not introducing a new method here and changing a name
at this point would be a
cloud-fan commented on code in PR #36576:
URL: https://github.com/apache/spark/pull/36576#discussion_r874598262
##
sql/core/src/test/scala/org/apache/spark/sql/SQLQuerySuite.scala:
##
@@ -147,6 +147,9 @@ class SQLQuerySuite extends QueryTest with
SharedSparkSession with
panbingkun commented on PR #36548:
URL: https://github.com/apache/spark/pull/36548#issuecomment-1128666500
@bozhang2820 @MaxGekk ping
--
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific
gengliangwang opened a new pull request, #36577:
URL: https://github.com/apache/spark/pull/36577
### What changes were proposed in this pull request?
1. Fix logical bugs in adding query contexts as references under codegen
mode.
linhongliu-db opened a new pull request, #36578:
URL: https://github.com/apache/spark/pull/36578
### What changes were proposed in this pull request?
Record the query text when executed with SQL API.
### Why are the changes needed?
* When executing a query using
linhongliu-db commented on PR #36578:
URL: https://github.com/apache/spark/pull/36578#issuecomment-1128693761
cc @cloud-fan, I think it should be useful to record the original SQL text
of a query.
--
This is an automated message from the Apache Git Service.
To respond to the message,
cloud-fan closed pull request #36572: [SPARK-36718][SQL][FOLLOWUP] Improve the
extract-only check in CollapseProject
URL: https://github.com/apache/spark/pull/36572
--
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL
cloud-fan commented on code in PR #36531:
URL: https://github.com/apache/spark/pull/36531#discussion_r874526787
##
sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/expressions/Cast.scala:
##
@@ -41,6 +40,113 @@ import org.apache.spark.unsafe.types.{CalendarInterval,
cloud-fan commented on code in PR #36531:
URL: https://github.com/apache/spark/pull/36531#discussion_r874525801
##
sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/analysis/TypeCoercion.scala:
##
@@ -772,15 +772,19 @@ abstract class TypeCoercionBase {
case e if
HyukjinKwon commented on code in PR #36576:
URL: https://github.com/apache/spark/pull/36576#discussion_r874624321
##
sql/core/src/test/scala/org/apache/spark/sql/SQLQuerySuite.scala:
##
@@ -147,6 +147,9 @@ class SQLQuerySuite extends QueryTest with
SharedSparkSession with
HyukjinKwon commented on code in PR #36576:
URL: https://github.com/apache/spark/pull/36576#discussion_r874632251
##
sql/core/src/test/scala/org/apache/spark/sql/BloomFilterAggregateQuerySuite.scala:
##
@@ -35,23 +34,26 @@ class BloomFilterAggregateQuerySuite extends QueryTest
cloud-fan commented on code in PR #36295:
URL: https://github.com/apache/spark/pull/36295#discussion_r874587041
##
sql/core/src/main/scala/org/apache/spark/sql/execution/datasources/v2/V2ScanRelationPushDown.scala:
##
@@ -44,6 +44,7 @@ object V2ScanRelationPushDown extends
cloud-fan commented on code in PR #36295:
URL: https://github.com/apache/spark/pull/36295#discussion_r874587459
##
sql/core/src/main/scala/org/apache/spark/sql/execution/datasources/v2/V2ScanRelationPushDown.scala:
##
@@ -44,6 +44,7 @@ object V2ScanRelationPushDown extends
physinet commented on code in PR #36545:
URL: https://github.com/apache/spark/pull/36545#discussion_r874781396
##
python/docs/source/migration_guide/pyspark_3.3_to_3.4.rst:
##
@@ -0,0 +1,23 @@
+.. Licensed to the Apache Software Foundation (ASF) under one
+or more
gengliangwang commented on PR #36582:
URL: https://github.com/apache/spark/pull/36582#issuecomment-1129038094
This should be the last one of query context fix when WSCG is not available.
--
This is an automated message from the Apache Git Service.
To respond to the message, please log on
gengliangwang opened a new pull request, #36582:
URL: https://github.com/apache/spark/pull/36582
### What changes were proposed in this pull request?
Similar to https://github.com/apache/spark/pull/36525, this PR provides
runtime error query context for the Average expression
srowen commented on PR #36496:
URL: https://github.com/apache/spark/pull/36496#issuecomment-1128869583
Hm, try tests again? I'm having trouble seeing the error. I thought it might
be MiMa, because you make a method private, but not sure that is it
--
This is an automated message from the
panbingkun opened a new pull request, #36580:
URL: https://github.com/apache/spark/pull/36580
### What changes were proposed in this pull request?
In the PR, I propose to use the MULTI_VALUE_SUBQUERY_ERROR error classes for
multiple rows from a subquery used as an expression.
###
HyukjinKwon commented on code in PR #36576:
URL: https://github.com/apache/spark/pull/36576#discussion_r874632251
##
sql/core/src/test/scala/org/apache/spark/sql/BloomFilterAggregateQuerySuite.scala:
##
@@ -35,23 +34,26 @@ class BloomFilterAggregateQuerySuite extends QueryTest
pan3793 commented on PR #36496:
URL: https://github.com/apache/spark/pull/36496#issuecomment-1128885052
Hmm, let me check the error message first
--
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to
srowen commented on PR #36567:
URL: https://github.com/apache/spark/pull/36567#issuecomment-112223
Merged to master
--
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
To
srowen commented on code in PR #36562:
URL: https://github.com/apache/spark/pull/36562#discussion_r874840258
##
sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/util/TimestampFormatter.scala:
##
@@ -52,6 +52,25 @@ sealed trait TimestampFormatter extends Serializable {
gengliangwang commented on code in PR #36562:
URL: https://github.com/apache/spark/pull/36562#discussion_r874864869
##
sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/util/TimestampFormatter.scala:
##
@@ -52,6 +52,25 @@ sealed trait TimestampFormatter extends
pan3793 commented on PR #36496:
URL: https://github.com/apache/spark/pull/36496#issuecomment-1128945680
Two jobs failed, hive slow test failed because of OOM, another is
pyspark(not familiar with python), re-triggered
--
This is an automated message from the Apache Git Service.
To
gengliangwang commented on PR #36577:
URL: https://github.com/apache/spark/pull/36577#issuecomment-1128945945
Merging to master/3.3
--
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific
gengliangwang closed pull request #36577: [SPARK-39208][SQL] Fix query context
bugs in decimal overflow under codegen mode
URL: https://github.com/apache/spark/pull/36577
--
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use
srowen commented on code in PR #36545:
URL: https://github.com/apache/spark/pull/36545#discussion_r875033677
##
python/pyspark/sql/session.py:
##
@@ -570,10 +570,20 @@ def _inferSchemaFromList(
if not data:
raise ValueError("can not infer schema from empty
neilagupta commented on PR #36441:
URL: https://github.com/apache/spark/pull/36441#issuecomment-1129086391
@AmplabJenkins any chance I could get someone with write access to review
this?
--
This is an automated message from the Apache Git Service.
To respond to the message, please log
HyukjinKwon commented on PR #36576:
URL: https://github.com/apache/spark/pull/36576#issuecomment-1128914759
Merged to master and branch-3.3.
--
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the
gengliangwang commented on code in PR #36562:
URL: https://github.com/apache/spark/pull/36562#discussion_r874872763
##
sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/util/TimestampFormatter.scala:
##
@@ -52,6 +52,25 @@ sealed trait TimestampFormatter extends
Yikun opened a new pull request, #36581:
URL: https://github.com/apache/spark/pull/36581
### What changes were proposed in this pull request?
Ensure sampling rows >= 2 to make sure apply's infer schema is accurate.
### Why are the changes needed?
GroupBy.apply infers schema
srowen closed pull request #36567: [SPARK-39196][CORE][SQL][K8S] replace
`getOrElse(null)` with `orNull`
URL: https://github.com/apache/spark/pull/36567
--
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go
srowen commented on PR #36529:
URL: https://github.com/apache/spark/pull/36529#issuecomment-1128890020
Merged to master
--
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
To
srowen closed pull request #36529: [SPARK-39102][CORE][SQL][DSTREAM] Add
checkstyle rules to disabled use of Guava's `Files.createTempDir()`
URL: https://github.com/apache/spark/pull/36529
--
This is an automated message from the Apache Git Service.
To respond to the message, please log on
pan3793 commented on PR #36496:
URL: https://github.com/apache/spark/pull/36496#issuecomment-1129007630
All tests past now
https://github.com/pan3793/spark/runs/6471801942?check_suite_focus=true
--
This is an automated message from the Apache Git Service.
To respond to the message,
physinet commented on code in PR #36545:
URL: https://github.com/apache/spark/pull/36545#discussion_r874983582
##
python/pyspark/sql/session.py:
##
@@ -570,10 +570,20 @@ def _inferSchemaFromList(
if not data:
raise ValueError("can not infer schema from
srowen commented on code in PR #36545:
URL: https://github.com/apache/spark/pull/36545#discussion_r874838643
##
python/pyspark/sql/session.py:
##
@@ -570,10 +570,20 @@ def _inferSchemaFromList(
if not data:
raise ValueError("can not infer schema from empty
HyukjinKwon closed pull request #36576: [SPARK-32268][SQL][TESTS][FOLLOW-UP]
Use function registry in the SparkSession
URL: https://github.com/apache/spark/pull/36576
--
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
Eugene-Mark commented on PR #36499:
URL: https://github.com/apache/spark/pull/36499#issuecomment-1129110003
@HyukjinKwon The [issue-38846
](https://issues.apache.org/jira/browse/SPARK-38846) shows that the Number type
of Teradata will lose its fractional part after loading to Spark. We
abellina commented on PR #36505:
URL: https://github.com/apache/spark/pull/36505#issuecomment-1129117318
Update on the SPARK-32290: SingleColumn Null Aware Anti Join Optimize
failure:
- The original test used a table in the subquery `testData2` which has no
nulls, so I added
dtenedor commented on PR #36583:
URL: https://github.com/apache/spark/pull/36583#issuecomment-1129125976
Note: this PR is based on https://github.com/apache/spark/pull/36501. The
additional changes comprise about 15 lines of code, in this commit:
physinet commented on code in PR #36545:
URL: https://github.com/apache/spark/pull/36545#discussion_r875099589
##
python/pyspark/sql/session.py:
##
@@ -570,10 +570,20 @@ def _inferSchemaFromList(
if not data:
raise ValueError("can not infer schema from
MaxGekk commented on code in PR #36553:
URL: https://github.com/apache/spark/pull/36553#discussion_r875133200
##
core/src/main/resources/error/error-classes.json:
##
@@ -22,8 +22,12 @@
"CANNOT_UP_CAST_DATATYPE" : {
"message" : [ "Cannot up cast from to
.\n" ]
},
-
physinet commented on code in PR #36545:
URL: https://github.com/apache/spark/pull/36545#discussion_r875142772
##
python/pyspark/sql/session.py:
##
@@ -570,10 +570,20 @@ def _inferSchemaFromList(
if not data:
raise ValueError("can not infer schema from
Eugene-Mark commented on PR #36499:
URL: https://github.com/apache/spark/pull/36499#issuecomment-1129099918
@srowen I'm also not a Teradata guy, just invokes Teradata's API from Spark
and found the issue. I didn't find the document explaining the issue at
Teradata side. I tried to print
abellina commented on PR #36505:
URL: https://github.com/apache/spark/pull/36505#issuecomment-1129153161
> All other queries in the test are passing, except for the negative case
for the multi-column support. It is commented out in my last patch (obviously
that's not the solution)
srowen commented on code in PR #36545:
URL: https://github.com/apache/spark/pull/36545#discussion_r875147010
##
python/pyspark/sql/session.py:
##
@@ -570,10 +570,20 @@ def _inferSchemaFromList(
if not data:
raise ValueError("can not infer schema from empty
MaxGekk commented on code in PR #36561:
URL: https://github.com/apache/spark/pull/36561#discussion_r875157818
##
sql/core/src/test/scala/org/apache/spark/sql/errors/QueryParsingErrorsSuite.scala:
##
@@ -642,4 +642,92 @@ class QueryParsingErrorsSuite extends QueryTest with
xinrong-databricks commented on code in PR #36560:
URL: https://github.com/apache/spark/pull/36560#discussion_r875138420
##
python/pyspark/pandas/tests/test_generic_functions.py:
##
@@ -150,8 +150,8 @@ def test_stat_functions(self):
vli-databricks opened a new pull request, #36584:
URL: https://github.com/apache/spark/pull/36584
### What changes were proposed in this pull request?
Adding implementation for ANY_VALUE aggregate function. During optimization
stage it is rewritten to `First` aggregate
MaxGekk commented on PR #36579:
URL: https://github.com/apache/spark/pull/36579#issuecomment-1129142133
@srielau @panbingkun Could you take a look at the PR, please.
--
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
srowen commented on code in PR #36545:
URL: https://github.com/apache/spark/pull/36545#discussion_r875104045
##
python/pyspark/sql/session.py:
##
@@ -570,10 +570,20 @@ def _inferSchemaFromList(
if not data:
raise ValueError("can not infer schema from empty
MaxGekk commented on PR #36553:
URL: https://github.com/apache/spark/pull/36553#issuecomment-1129174496
cc @srielau
--
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
To
dtenedor commented on PR #36501:
URL: https://github.com/apache/spark/pull/36501#issuecomment-1129093762
@HyukjinKwon I fixed the bad sync, this is ready to merge now at your
convenience.
--
This is an automated message from the Apache Git Service.
To respond to the message, please log
srowen commented on PR #36499:
URL: https://github.com/apache/spark/pull/36499#issuecomment-1129130839
OK, I just wonder if this is specific to Teradata, or whether it can be
changed elsewhere higher up in the abstraction layers.
But you're saying the scale/precision info is lost in
dtenedor opened a new pull request, #36583:
URL: https://github.com/apache/spark/pull/36583
### What changes were proposed in this pull request?
Support JSON scans when the table schema has associated DEFAULT column
values.
Example:
```
create table t(i int) using
xinrong-databricks commented on PR #36547:
URL: https://github.com/apache/spark/pull/36547#issuecomment-1129166446
Rebased master to retrigger irrelevant failed test. No new changes after
review.
--
This is an automated message from the Apache Git Service.
To respond to the message,
xinrong-databricks commented on PR #36547:
URL: https://github.com/apache/spark/pull/36547#issuecomment-1129178343
> Boolean cast not only is not going to cover all types, but also yield
different results in some cases
Would you give an example in which case we may diverge from
zero323 commented on PR #36547:
URL: https://github.com/apache/spark/pull/36547#issuecomment-1129228724
> Would you give an example in which case we may diverge from pandas? I
Sure thing @xinrong-databricks. Sorry for being enigmatic before. So, very
simple case would be something
srowen commented on PR #36496:
URL: https://github.com/apache/spark/pull/36496#issuecomment-1129416530
Merged to master/3.3/3.2
--
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific
dtenedor commented on PR #36583:
URL: https://github.com/apache/spark/pull/36583#issuecomment-1129351769
> Is this
[[[SPARK-38067](https://issues.apache.org/jira/browse/SPARK-38067)][PYTHON]
Preserve None values when saved to
HyukjinKwon commented on PR #36581:
URL: https://github.com/apache/spark/pull/36581#issuecomment-1129428282
Merged to master.
--
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific
amaliujia commented on code in PR #36586:
URL: https://github.com/apache/spark/pull/36586#discussion_r875377592
##
sql/core/src/main/scala/org/apache/spark/sql/internal/CatalogImpl.scala:
##
@@ -204,8 +211,12 @@ class CatalogImpl(sparkSession: SparkSession) extends
Catalog {
cloud-fan commented on code in PR #36586:
URL: https://github.com/apache/spark/pull/36586#discussion_r875383064
##
sql/catalyst/src/test/scala/org/apache/spark/sql/catalyst/catalog/ExternalCatalogSuite.scala:
##
@@ -1025,9 +1025,14 @@ abstract class CatalogTestUtils {
def
cloud-fan commented on code in PR #36586:
URL: https://github.com/apache/spark/pull/36586#discussion_r875383670
##
sql/core/src/main/scala/org/apache/spark/sql/execution/datasources/v2/ShowTablesExec.scala:
##
@@ -38,12 +38,14 @@ case class ShowTablesExec(
val rows = new
HyukjinKwon opened a new pull request, #36587:
URL: https://github.com/apache/spark/pull/36587
### What changes were proposed in this pull request?
This PR proposes to reduce the number of Py4J calls at
`pyspark.sql.utils.is_timestamp_ntz_preferred` by having a single method to
gengliangwang commented on PR #36562:
URL: https://github.com/apache/spark/pull/36562#issuecomment-1129515141
Merging to master/3.3
--
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific
1 - 100 of 156 matches
Mail list logo