zhengruifeng commented on code in PR #36246:
URL: https://github.com/apache/spark/pull/36246#discussion_r855817585
##
python/pyspark/pandas/series.py:
##
@@ -2209,15 +2219,43 @@ def _interpolate(
) * null_index_forward + last_non_null_forward
fill_cond = ~F.i
panbingkun commented on code in PR #36314:
URL: https://github.com/apache/spark/pull/36314#discussion_r855814432
##
sql/core/src/test/scala/org/apache/spark/sql/errors/QueryExecutionAnsiErrorsSuite.scala:
##
@@ -77,4 +77,23 @@ class QueryExecutionAnsiErrorsSuite extends QueryTes
xiuzhu9527 commented on PR #36217:
URL: https://github.com/apache/spark/pull/36217#issuecomment-1106063878
> Why not?
spark-tags-tests.jar is used in maven test phase.
--
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub an
MaxGekk commented on PR #35269:
URL: https://github.com/apache/spark/pull/35269#issuecomment-1106047562
FYI, the feature is in the allow list for Spark 3.3, and in fact 3.3 is
waiting for only this PR. @cloud-fan @dtenedor @amaliujia @beliefer How long
could it take to be ready for merging
MaxGekk commented on PR #36232:
URL: https://github.com/apache/spark/pull/36232#issuecomment-1106041088
@panbingkun Could you resolve conflicts, here.
--
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go
MaxGekk commented on code in PR #36280:
URL: https://github.com/apache/spark/pull/36280#discussion_r855792086
##
sql/core/src/test/scala/org/apache/spark/sql/errors/QueryCompilationErrorsSuite.scala:
##
@@ -362,6 +362,45 @@ class QueryCompilationErrorsSuite extends QueryTest wit
MaxGekk commented on code in PR #36284:
URL: https://github.com/apache/spark/pull/36284#discussion_r855791244
##
sql/core/src/test/scala/org/apache/spark/sql/errors/QueryCompilationErrorsSuite.scala:
##
@@ -362,6 +362,19 @@ class QueryCompilationErrorsSuite extends QueryTest wit
MaxGekk commented on code in PR #36298:
URL: https://github.com/apache/spark/pull/36298#discussion_r855790900
##
sql/core/src/test/scala/org/apache/spark/sql/errors/QueryCompilationErrorsSuite.scala:
##
@@ -364,6 +364,21 @@ class QueryCompilationErrorsSuite extends QueryTest wit
MaxGekk commented on code in PR #36314:
URL: https://github.com/apache/spark/pull/36314#discussion_r855790606
##
sql/core/src/test/scala/org/apache/spark/sql/errors/QueryExecutionAnsiErrorsSuite.scala:
##
@@ -77,4 +77,23 @@ class QueryExecutionAnsiErrorsSuite extends QueryTest w
MaxGekk commented on code in PR #36320:
URL: https://github.com/apache/spark/pull/36320#discussion_r855790359
##
sql/core/src/test/scala/org/apache/spark/sql/errors/QueryExecutionErrorsSuite.scala:
##
@@ -373,4 +373,29 @@ class QueryExecutionErrorsSuite extends QueryTest
as
gengliangwang closed pull request #36316: [SPARK-38813][3.3][SQL][FOLLOWUP]
Improve the analysis check for TimestampNTZ output
URL: https://github.com/apache/spark/pull/36316
--
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and us
HyukjinKwon closed pull request #36318: [SPARK-38994][DOCS] Add an Python
example of StreamingQueryListener
URL: https://github.com/apache/spark/pull/36318
--
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to
gengliangwang commented on PR #36316:
URL: https://github.com/apache/spark/pull/36316#issuecomment-1106027911
@cloud-fan @dongjoon-hyun @ueshin Thanks for the review
Merging to 3.3
--
This is an automated message from the Apache Git Service.
To respond to the message, please log on to G
HyukjinKwon commented on PR #36318:
URL: https://github.com/apache/spark/pull/36318#issuecomment-1106027550
Merged to master.
--
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
MaxGekk closed pull request #36287: [SPARK-38986][SQL] Prepend error class tag
to error messages
URL: https://github.com/apache/spark/pull/36287
--
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the s
MaxGekk commented on PR #36287:
URL: https://github.com/apache/spark/pull/36287#issuecomment-1106015680
GA passed. Merging to master.
Thank you, @cloud-fan and @HyukjinKwon for review.
--
This is an automated message from the Apache Git Service.
To respond to the message, please log on
MaxGekk commented on code in PR #36287:
URL: https://github.com/apache/spark/pull/36287#discussion_r855770217
##
core/src/main/scala/org/apache/spark/ErrorInfo.scala:
##
@@ -58,7 +58,8 @@ private[spark] object SparkThrowableHelper {
def getMessage(errorClass: String, messageP
lvshaokang opened a new pull request, #36320:
URL: https://github.com/apache/spark/pull/36320
### What changes were proposed in this pull request?
I add a test case for the error class INCOMPARABLE_PIVOT_COLUMN in the
QueryExecutionErrorsSuite.
### Why are the changes n
beliefer opened a new pull request, #36319:
URL: https://github.com/apache/spark/pull/36319
### What changes were proposed in this pull request?
https://github.com/apache/spark/pull/35975 supported ANSI SQL: result offset
clause.
We make some check for offset in `CheckAnalysis`.
The
HyukjinKwon commented on PR #36318:
URL: https://github.com/apache/spark/pull/36318#issuecomment-110633
cc @HeartSaVioR
--
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
HyukjinKwon opened a new pull request, #36318:
URL: https://github.com/apache/spark/pull/36318
### What changes were proposed in this pull request?
This PR proposes to add an example of `StreamingQueryListener` in Python
addd in SPARK-38759.
### Why are the changes needed?
itholic commented on PR #36306:
URL: https://github.com/apache/spark/pull/36306#issuecomment-1105993698
Maybe the JIRA number is incorrect ??
Seems like it should be SPARK-38989.
--
This is an automated message from the Apache Git Service.
To respond to the message, please log on to
JoshRosen commented on code in PR #36315:
URL: https://github.com/apache/spark/pull/36315#discussion_r855751881
##
core/src/main/scala/org/apache/spark/security/ShellBasedGroupsMappingProvider.scala:
##
@@ -36,10 +36,12 @@ private[spark] class ShellBasedGroupsMappingProvider
ex
HyukjinKwon commented on code in PR #36315:
URL: https://github.com/apache/spark/pull/36315#discussion_r855751736
##
core/src/main/scala/org/apache/spark/security/ShellBasedGroupsMappingProvider.scala:
##
@@ -36,10 +36,12 @@ private[spark] class ShellBasedGroupsMappingProvider
JoshRosen commented on code in PR #36315:
URL: https://github.com/apache/spark/pull/36315#discussion_r855748988
##
core/src/main/scala/org/apache/spark/security/ShellBasedGroupsMappingProvider.scala:
##
@@ -36,10 +36,12 @@ private[spark] class ShellBasedGroupsMappingProvider
ex
JoshRosen commented on code in PR #36315:
URL: https://github.com/apache/spark/pull/36315#discussion_r855748988
##
core/src/main/scala/org/apache/spark/security/ShellBasedGroupsMappingProvider.scala:
##
@@ -36,10 +36,12 @@ private[spark] class ShellBasedGroupsMappingProvider
ex
dongjoon-hyun commented on PR #36316:
URL: https://github.com/apache/spark/pull/36316#issuecomment-1105983522
Got it, @gengliangwang . Never mind~ :)
--
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go
gengliangwang commented on PR #36316:
URL: https://github.com/apache/spark/pull/36316#issuecomment-1105982873
@dongjoon-hyun Yes I would love to.
There is a check `!Utils.isTesting`. If I temporarily disable it in one test
case, will it be a potential issue for running tests in parallel?
srowen commented on PR #36217:
URL: https://github.com/apache/spark/pull/36217#issuecomment-1105981447
Why not?
--
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
To unsubscri
JoshRosen commented on code in PR #36315:
URL: https://github.com/apache/spark/pull/36315#discussion_r855748988
##
core/src/main/scala/org/apache/spark/security/ShellBasedGroupsMappingProvider.scala:
##
@@ -36,10 +36,12 @@ private[spark] class ShellBasedGroupsMappingProvider
ex
JoshRosen commented on code in PR #36315:
URL: https://github.com/apache/spark/pull/36315#discussion_r855748988
##
core/src/main/scala/org/apache/spark/security/ShellBasedGroupsMappingProvider.scala:
##
@@ -36,10 +36,12 @@ private[spark] class ShellBasedGroupsMappingProvider
ex
gengliangwang commented on code in PR #36315:
URL: https://github.com/apache/spark/pull/36315#discussion_r855746240
##
core/src/main/scala/org/apache/spark/security/ShellBasedGroupsMappingProvider.scala:
##
@@ -38,8 +38,10 @@ private[spark] class ShellBasedGroupsMappingProvider
HyukjinKwon commented on code in PR #36315:
URL: https://github.com/apache/spark/pull/36315#discussion_r855748245
##
core/src/main/scala/org/apache/spark/security/ShellBasedGroupsMappingProvider.scala:
##
@@ -38,8 +38,10 @@ private[spark] class ShellBasedGroupsMappingProvider ex
lvshaokang closed pull request #36297: [SPARK-38732][SQL][TESTS] Test the error
class: INCOMPARABLE_PIVOT_COLUMN
URL: https://github.com/apache/spark/pull/36297
--
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL abov
gengliangwang commented on code in PR #36315:
URL: https://github.com/apache/spark/pull/36315#discussion_r855746240
##
core/src/main/scala/org/apache/spark/security/ShellBasedGroupsMappingProvider.scala:
##
@@ -38,8 +38,10 @@ private[spark] class ShellBasedGroupsMappingProvider
xiuzhu9527 commented on PR #36217:
URL: https://github.com/apache/spark/pull/36217#issuecomment-1105973624
> This does not explain why?
> > This does not explain why?
> > Thank you very much for your reply!After spark build is completed, the
$spark_project_home/assembly/tar
xiuzhu9527 commented on PR #36217:
URL: https://github.com/apache/spark/pull/36217#issuecomment-1105973109
> This does not explain why?
Thank you very much for your reply!After spark build is completed, the
$spark_project_home/assembly/target/scala-2.11/jars directory contains
HyukjinKwon commented on PR #36312:
URL: https://github.com/apache/spark/pull/36312#issuecomment-1105968096
Merged to master, branch-3.3, branch-3.2, branch-3.1 and branch-3.0.
--
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub a
HyukjinKwon closed pull request #36312: [SPARK-38990][SQL] Avoid
`NullPointerException` when evaluating date_trunc/trunc format as a bound
reference
URL: https://github.com/apache/spark/pull/36312
--
This is an automated message from the Apache Git Service.
To respond to the message, please
xiuzhu9527 commented on PR #36217:
URL: https://github.com/apache/spark/pull/36217#issuecomment-1105965230
> @dongjoon-hyun Could you take a look? thanks!
--
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL
cloud-fan closed pull request #36291: [SPARK-38974][SQL] Filter registered
functions with a given database name in list functions
URL: https://github.com/apache/spark/pull/36291
--
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and
cloud-fan commented on PR #36291:
URL: https://github.com/apache/spark/pull/36291#issuecomment-1105964564
thanks, merging to master/3.3!
--
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specif
zhengruifeng opened a new pull request, #36317:
URL: https://github.com/apache/spark/pull/36317
### What changes were proposed in this pull request?
Impl DataFrame.boxplot and DataFrame.plot.box
### Why are the changes needed?
to increase pandas API coverage in PySpark
cloud-fan commented on code in PR #36303:
URL: https://github.com/apache/spark/pull/36303#discussion_r855736358
##
sql/core/src/test/scala/org/apache/spark/sql/execution/datasources/SchemaPruningSuite.scala:
##
@@ -935,4 +935,106 @@ abstract class SchemaPruningSuite
.coun
gengliangwang opened a new pull request, #36316:
URL: https://github.com/apache/spark/pull/36316
### What changes were proposed in this pull request?
In https://github.com/apache/spark/pull/36094, a check for failing
TimestampNTZ output is added.
However, if there is an unr
aray commented on PR #36150:
URL: https://github.com/apache/spark/pull/36150#issuecomment-1105961462
@EnricoMi thanks for the PR. This has been a TODO for years now since I
added pivot. If you want to implement this with `stack` you can just use the
expression directly, no need to add a fun
xiuzhu9527 commented on PR #36217:
URL: https://github.com/apache/spark/pull/36217#issuecomment-1105961386
> @dongjoon-hyun Could you take a look? thanks!
--
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL
xiuzhu9527 closed pull request #36217: [BUILD] When building spark project,
remove spark-tags-tests.jar from…
URL: https://github.com/apache/spark/pull/36217
--
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above t
HyukjinKwon commented on code in PR #36296:
URL: https://github.com/apache/spark/pull/36296#discussion_r855735028
##
sql/core/src/main/scala/org/apache/spark/sql/execution/datasources/orc/OrcUtils.scala:
##
@@ -224,8 +224,10 @@ object OrcUtils extends Logging {
// the p
cloud-fan commented on PR #36072:
URL: https://github.com/apache/spark/pull/36072#issuecomment-1105959980
thanks, merging to master/3.3!
--
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specif
HyukjinKwon closed pull request #36308: [SPARK-38581][PYTHON][DOCS][3.3] List
of supported pandas APIs for pandas-on-Spark docs
URL: https://github.com/apache/spark/pull/36308
--
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and u
cloud-fan closed pull request #36072: [SPARK-38666][SQL] Add missing aggregate
filter checks
URL: https://github.com/apache/spark/pull/36072
--
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the speci
HyukjinKwon commented on PR #36315:
URL: https://github.com/apache/spark/pull/36315#issuecomment-1105959105
cc @JoshRosen @gengliangwang mind taking a look please?
--
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
UR
HyukjinKwon opened a new pull request, #36315:
URL: https://github.com/apache/spark/pull/36315
### What changes were proposed in this pull request?
This PR proposes to avoid using `bash -c` in
`ShellBasedGroupsMappingProvider`. This could allow users a command injection.
### Wh
cloud-fan commented on code in PR #36230:
URL: https://github.com/apache/spark/pull/36230#discussion_r855731336
##
sql/catalyst/src/test/scala/org/apache/spark/sql/catalyst/optimizer/OuterJoinEliminationSuite.scala:
##
@@ -252,4 +254,18 @@ class OuterJoinEliminationSuite extends
gengliangwang commented on code in PR #36094:
URL: https://github.com/apache/spark/pull/36094#discussion_r855731255
##
sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/analysis/CheckAnalysis.scala:
##
@@ -157,6 +158,10 @@ trait CheckAnalysis extends PredicateHelper with
cloud-fan commented on code in PR #36230:
URL: https://github.com/apache/spark/pull/36230#discussion_r855731223
##
sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/optimizer/joins.scala:
##
@@ -144,8 +144,17 @@ object EliminateOuterJoin extends Rule[LogicalPlan] with
P
cloud-fan commented on code in PR #36094:
URL: https://github.com/apache/spark/pull/36094#discussion_r855730722
##
sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/analysis/CheckAnalysis.scala:
##
@@ -157,6 +158,10 @@ trait CheckAnalysis extends PredicateHelper with
Lo
cloud-fan commented on code in PR #36094:
URL: https://github.com/apache/spark/pull/36094#discussion_r855730647
##
sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/analysis/CheckAnalysis.scala:
##
@@ -157,6 +158,10 @@ trait CheckAnalysis extends PredicateHelper with
Lo
srowen commented on PR #36217:
URL: https://github.com/apache/spark/pull/36217#issuecomment-1105953110
This does not explain why?
--
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comm
itholic commented on code in PR #36306:
URL: https://github.com/apache/spark/pull/36306#discussion_r855727915
##
python/pyspark/pandas/frame.py:
##
@@ -8689,7 +8702,13 @@ def sample(
sdf = self._internal.resolved_copy.spark_frame.sample(
withReplacement=rep
PavithraRamachandran commented on PR #36278:
URL: https://github.com/apache/spark/pull/36278#issuecomment-1105950450
@martin-g @LuciferYang could u help review this?
--
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
xiuzhu9527 commented on PR #36217:
URL: https://github.com/apache/spark/pull/36217#issuecomment-1105948761
cc @dongjoon-hyun @HyukjinKwon @srowen
--
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to
HyukjinKwon closed pull request #36215: [SPARK-38938][PYTHON] Implement
`inplace` and `columns` parameters of `Series.drop`
URL: https://github.com/apache/spark/pull/36215
--
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use t
HyukjinKwon commented on PR #36215:
URL: https://github.com/apache/spark/pull/36215#issuecomment-1105948369
Merged to master.
--
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
HyukjinKwon closed pull request #36266: [SPARK-38952][PYTHON] Implement
`numeric_only` of `GroupBy.first` and `GroupBy.last`
URL: https://github.com/apache/spark/pull/36266
--
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use
HyukjinKwon commented on PR #36266:
URL: https://github.com/apache/spark/pull/36266#issuecomment-1105948075
Merged to master.
--
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
HyukjinKwon closed pull request #36294: [SPARK-38955][SQL] Disable lineSep
option in 'from_csv' and 'schema_of_csv'
URL: https://github.com/apache/spark/pull/36294
--
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL a
HyukjinKwon commented on PR #36294:
URL: https://github.com/apache/spark/pull/36294#issuecomment-1105947034
I will get this in first since 3.3 RC is coming soon.
Merged to master and branch-3.3.
--
This is an automated message from the Apache Git Service.
To respond to the message,
HyukjinKwon commented on PR #36294:
URL: https://github.com/apache/spark/pull/36294#issuecomment-1105945488
Yeah. I actually think we should check all options there and document them.
Actually, we might even throw an exception too like `parseMode` case but it
might be too breaking.
--
Th
HyukjinKwon commented on code in PR #36287:
URL: https://github.com/apache/spark/pull/36287#discussion_r855722459
##
core/src/main/scala/org/apache/spark/ErrorInfo.scala:
##
@@ -58,7 +58,8 @@ private[spark] object SparkThrowableHelper {
def getMessage(errorClass: String, mess
panbingkun opened a new pull request, #36314:
URL: https://github.com/apache/spark/pull/36314
## What changes were proposed in this pull request?
This pr aims to add one test for the error class INVALID_ARRAY_INDEX &
INVALID_ARRAY_INDEX_IN_ELEMENT_AT to QueryExecutionAnsiErrorsSuite.
panbingkun closed pull request #36313: [SPARK-38734][SQL][TESTS] Test the error
classes: INVALID_ARRAY_INDEX & INVALID_ARRAY_INDEX_IN_ELEMENT_AT
URL: https://github.com/apache/spark/pull/36313
--
This is an automated message from the Apache Git Service.
To respond to the message, please log o
HyukjinKwon commented on PR #36308:
URL: https://github.com/apache/spark/pull/36308#issuecomment-1105941459
Merged to banch-3.3.
--
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comme
panbingkun opened a new pull request, #36313:
URL: https://github.com/apache/spark/pull/36313
## What changes were proposed in this pull request?
This pr aims to add one test for the error class INVALID_ARRAY_INDEX &
INVALID_ARRAY_INDEX_IN_ELEMENT_AT to QueryExecutionAnsiErrorsSuite.
HyukjinKwon commented on PR #36038:
URL: https://github.com/apache/spark/pull/36038#issuecomment-1105940874
oh yeah. will take a look. thanks for pointing this out!
--
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
U
dongjoon-hyun commented on PR #36310:
URL: https://github.com/apache/spark/pull/36310#issuecomment-1105940185
Merged to all live branches (master/3.3/3.2/3.1).
--
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL ab
dongjoon-hyun closed pull request #36310: [MINOR][DOCS] Also remove Google
Analytics from Spark release docs, per ASF policy
URL: https://github.com/apache/spark/pull/36310
--
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use
bersprockets opened a new pull request, #36312:
URL: https://github.com/apache/spark/pull/36312
### What changes were proposed in this pull request?
Change `TruncInstant.evalHelper` to pass the input row to `format.eval` when
`format` is a not a literal (and therefore might be a bound
c21 commented on PR #36311:
URL: https://github.com/apache/spark/pull/36311#issuecomment-1105909529
@cloud-fan and @huaxingao could you help take a look when you have time?
Thanks.
--
This is an automated message from the Apache Git Service.
To respond to the message, please log on to Git
c21 opened a new pull request, #36311:
URL: https://github.com/apache/spark/pull/36311
### What changes were proposed in this pull request?
This is a followup per comment in
https://issues.apache.org/jira/browse/SPARK-34960, to improve the documentation
for data source v2 agg
zhengruifeng commented on code in PR #36205:
URL: https://github.com/apache/spark/pull/36205#discussion_r855689831
##
python/pyspark/pandas/frame.py:
##
@@ -1310,6 +1310,137 @@ def corr(self, method: str = "pearson") -> "DataFrame":
"""
return cast(DataFrame, p
srowen opened a new pull request, #36310:
URL: https://github.com/apache/spark/pull/36310
### What changes were proposed in this pull request?
Remove Google Analytics from Spark release docs.
See also https://github.com/apache/spark-website/pull/384
### Why are the changes ne
github-actions[bot] commented on PR #34995:
URL: https://github.com/apache/spark/pull/34995#issuecomment-1105878086
We're closing this PR because it hasn't been updated in a while. This isn't
a judgement on the merit of the PR in any way. It's just a way of keeping the
PR queue manageable.
github-actions[bot] commented on PR #35076:
URL: https://github.com/apache/spark/pull/35076#issuecomment-1105878067
We're closing this PR because it hasn't been updated in a while. This isn't
a judgement on the merit of the PR in any way. It's just a way of keeping the
PR queue manageable.
github-actions[bot] closed pull request #35140: [SPARK-37829][SQL]
DataFrame.joinWith should return null rows for missing values
URL: https://github.com/apache/spark/pull/35140
--
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and
xinrong-databricks opened a new pull request, #36309:
URL: https://github.com/apache/spark/pull/36309
### What changes were proposed in this pull request?
### Why are the changes needed?
### Does this PR introduce _any_ user-facing change?
itholic commented on PR #36083:
URL: https://github.com/apache/spark/pull/36083#issuecomment-1105857345
@HyukjinKwon Just created at https://github.com/apache/spark/pull/36308 !
--
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub
itholic opened a new pull request, #36308:
URL: https://github.com/apache/spark/pull/36308
### What changes were proposed in this pull request?
This PR proposes to add new page named "Supported pandas APIs" for
pandas-on-Spark documents.
This is cherry-pick from
https://github
srielau opened a new pull request, #36307:
URL: https://github.com/apache/spark/pull/36307
### What changes were proposed in this pull request?
### Why are the changes needed?
### Does this PR introduce _any_ user-facing change?
### How was
itholic commented on PR #36267:
URL: https://github.com/apache/spark/pull/36267#issuecomment-1105833578
Would mind adding the screen-capture of the some part of page so that we can
easily verify the page is rendered properly ??
--
This is an automated message from the Apache Git Service.
itholic commented on PR #36205:
URL: https://github.com/apache/spark/pull/36205#issuecomment-1105832399
Also same as
https://github.com/apache/spark/pull/36246#issuecomment-1105830519.
--
This is an automated message from the Apache Git Service.
To respond to the message, please log on to
itholic commented on PR #36246:
URL: https://github.com/apache/spark/pull/36246#issuecomment-1105830519
Could you also update
python/docs/source/user_guide/pandas_on_spark/supported_pandas_api.rst ?
We should keep this list up-to-date manually for now when adding the new API
or param
huaxingao commented on code in PR #36303:
URL: https://github.com/apache/spark/pull/36303#discussion_r855641933
##
sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/expressions/SchemaPruning.scala:
##
@@ -152,6 +152,10 @@ object SchemaPruning extends SQLConfHelper {
sadikovi commented on PR #36158:
URL: https://github.com/apache/spark/pull/36158#issuecomment-1105825764
@gengliangwang @sunchao thanks for the review! I addressed the comments.
--
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub
sadikovi commented on code in PR #36158:
URL: https://github.com/apache/spark/pull/36158#discussion_r855640236
##
sql/catalyst/src/main/scala/org/apache/spark/sql/internal/SQLConf.scala:
##
@@ -1061,6 +1061,16 @@ object SQLConf {
.booleanConf
.createWithDefault(fal
sadikovi commented on code in PR #36158:
URL: https://github.com/apache/spark/pull/36158#discussion_r855639984
##
sql/catalyst/src/main/scala/org/apache/spark/sql/internal/SQLConf.scala:
##
@@ -1061,6 +1061,16 @@ object SQLConf {
.booleanConf
.createWithDefault(fal
sadikovi commented on code in PR #36158:
URL: https://github.com/apache/spark/pull/36158#discussion_r855639984
##
sql/catalyst/src/main/scala/org/apache/spark/sql/internal/SQLConf.scala:
##
@@ -1061,6 +1061,16 @@ object SQLConf {
.booleanConf
.createWithDefault(fal
huaxingao commented on code in PR #36303:
URL: https://github.com/apache/spark/pull/36303#discussion_r855637167
##
sql/core/src/test/scala/org/apache/spark/sql/execution/datasources/SchemaPruningSuite.scala:
##
@@ -935,4 +935,106 @@ abstract class SchemaPruningSuite
.coun
huaxingao commented on code in PR #36303:
URL: https://github.com/apache/spark/pull/36303#discussion_r855636865
##
sql/core/src/test/scala/org/apache/spark/sql/execution/datasources/SchemaPruningSuite.scala:
##
@@ -935,4 +935,106 @@ abstract class SchemaPruningSuite
.coun
1 - 100 of 204 matches
Mail list logo