Github user JoshRosen commented on the pull request:
https://github.com/apache/spark/pull/4155#issuecomment-71227419
Woohoo, looks like this is passing tests! The earlier failure was due to a
known flaky streaming test.
---
If your project is set up for it, you can reply to this
Github user AmplabJenkins commented on the pull request:
https://github.com/apache/spark/pull/4182#issuecomment-71230268
Test PASSed.
Refer to this link for build results (access rights to CI server needed):
Github user AmplabJenkins commented on the pull request:
https://github.com/apache/spark/pull/4183#issuecomment-71232329
Test PASSed.
Refer to this link for build results (access rights to CI server needed):
Github user SparkQA commented on the pull request:
https://github.com/apache/spark/pull/4183#issuecomment-71232323
[Test build #26029 has
finished](https://amplab.cs.berkeley.edu/jenkins/job/SparkPullRequestBuilder/26029/consoleFull)
for PR 4183 at commit
Github user vanzin commented on the pull request:
https://github.com/apache/spark/pull/3658#issuecomment-71236254
Ping.
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled
Github user SparkQA commented on the pull request:
https://github.com/apache/spark/pull/3916#issuecomment-71237900
[Test build #26031 has
started](https://amplab.cs.berkeley.edu/jenkins/job/SparkPullRequestBuilder/26031/consoleFull)
for PR 3916 at commit
GitHub user hhbyyh opened a pull request:
https://github.com/apache/spark/pull/4183
[SPARK-5384][mllib] Vectors.sqdist return inconsistent result for
sparse/dense vectors when the vectors have different lengths
JIRA issue: https://issues.apache.org/jira/browse/SPARK-5384
Github user SparkQA commented on the pull request:
https://github.com/apache/spark/pull/4183#issuecomment-71220720
[Test build #26029 has
started](https://amplab.cs.berkeley.edu/jenkins/job/SparkPullRequestBuilder/26029/consoleFull)
for PR 4183 at commit
Github user SparkQA commented on the pull request:
https://github.com/apache/spark/pull/4155#issuecomment-71225754
[Test build #26027 has
finished](https://amplab.cs.berkeley.edu/jenkins/job/SparkPullRequestBuilder/26027/consoleFull)
for PR 4155 at commit
Github user SparkQA commented on the pull request:
https://github.com/apache/spark/pull/4178#issuecomment-71229958
[Test build #26030 has
started](https://amplab.cs.berkeley.edu/jenkins/job/SparkPullRequestBuilder/26030/consoleFull)
for PR 4178 at commit
Github user AmplabJenkins commented on the pull request:
https://github.com/apache/spark/pull/4155#issuecomment-71225766
Test PASSed.
Refer to this link for build results (access rights to CI server needed):
Github user SparkQA commented on the pull request:
https://github.com/apache/spark/pull/3884#issuecomment-71225070
[Test build #26026 has
finished](https://amplab.cs.berkeley.edu/jenkins/job/SparkPullRequestBuilder/26026/consoleFull)
for PR 3884 at commit
Github user AmplabJenkins commented on the pull request:
https://github.com/apache/spark/pull/3884#issuecomment-71225077
Test PASSed.
Refer to this link for build results (access rights to CI server needed):
Github user SparkQA commented on the pull request:
https://github.com/apache/spark/pull/4182#issuecomment-71230259
[Test build #26028 has
finished](https://amplab.cs.berkeley.edu/jenkins/job/SparkPullRequestBuilder/26028/consoleFull)
for PR 4182 at commit
Github user vanzin commented on the pull request:
https://github.com/apache/spark/pull/4082#issuecomment-71248520
@ksakellis
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
Github user SparkQA commented on the pull request:
https://github.com/apache/spark/pull/4178#issuecomment-71241536
[Test build #26030 has
finished](https://amplab.cs.berkeley.edu/jenkins/job/SparkPullRequestBuilder/26030/consoleFull)
for PR 4178 at commit
Github user vanzin commented on the pull request:
https://github.com/apache/spark/pull/3884#issuecomment-71244412
Scala changes look ok to me; I'm not super familiar with the pyspark
internals, but the check on `rdd.py` surprised me because I thought RDDs were
actually serialized, at
Github user vanzin commented on a diff in the pull request:
https://github.com/apache/spark/pull/4155#discussion_r23473037
--- Diff:
core/src/main/scala/org/apache/spark/scheduler/OutputCommitCoordinator.scala ---
@@ -0,0 +1,178 @@
+/*
+ * Licensed to the Apache Software
Github user vanzin commented on the pull request:
https://github.com/apache/spark/pull/4155#issuecomment-71252591
I had this (unbased) notion that tasks knew whether they were speculative
or not, and thus the non-speculative ones would be able to avoid this extra
hop to the driver
Github user SparkQA commented on the pull request:
https://github.com/apache/spark/pull/3916#issuecomment-71255179
[Test build #26031 has
finished](https://amplab.cs.berkeley.edu/jenkins/job/SparkPullRequestBuilder/26031/consoleFull)
for PR 3916 at commit
Github user AmplabJenkins commented on the pull request:
https://github.com/apache/spark/pull/4173#issuecomment-71259112
Test FAILed.
Refer to this link for build results (access rights to CI server needed):
Github user SparkQA commented on the pull request:
https://github.com/apache/spark/pull/4173#issuecomment-71259100
[Test build #26033 has
finished](https://amplab.cs.berkeley.edu/jenkins/job/SparkPullRequestBuilder/26033/consoleFull)
for PR 4173 at commit
Github user mccheah commented on a diff in the pull request:
https://github.com/apache/spark/pull/4155#discussion_r23478673
--- Diff: core/src/main/scala/org/apache/spark/SparkHadoopWriter.scala ---
@@ -106,18 +107,25 @@ class SparkHadoopWriter(@transient jobConf: JobConf)
Github user vanzin commented on the pull request:
https://github.com/apache/spark/pull/4155#issuecomment-71262056
We do actually need the processing to be single threaded, as trying to
coordinate synchronization on the centralized arbitration logic is a bit of a
nightmare.
Github user SparkQA commented on the pull request:
https://github.com/apache/spark/pull/4139#issuecomment-71246755
[Test build #26035 has
started](https://amplab.cs.berkeley.edu/jenkins/job/SparkPullRequestBuilder/26035/consoleFull)
for PR 4139 at commit
Github user AmplabJenkins commented on the pull request:
https://github.com/apache/spark/pull/4178#issuecomment-71241547
Test PASSed.
Refer to this link for build results (access rights to CI server needed):
Github user vanzin commented on a diff in the pull request:
https://github.com/apache/spark/pull/4155#discussion_r23473132
--- Diff:
core/src/main/scala/org/apache/spark/scheduler/OutputCommitCoordinator.scala ---
@@ -0,0 +1,178 @@
+/*
+ * Licensed to the Apache Software
Github user vanzin commented on a diff in the pull request:
https://github.com/apache/spark/pull/4155#discussion_r23478213
--- Diff: core/src/main/scala/org/apache/spark/scheduler/DAGScheduler.scala
---
@@ -808,6 +810,7 @@ class DAGScheduler(
// will be posted, which
Github user vanzin commented on a diff in the pull request:
https://github.com/apache/spark/pull/4155#discussion_r23478509
--- Diff: core/src/main/scala/org/apache/spark/SparkHadoopWriter.scala ---
@@ -106,18 +107,25 @@ class SparkHadoopWriter(@transient jobConf: JobConf)
Github user AmplabJenkins commented on the pull request:
https://github.com/apache/spark/pull/4173#issuecomment-71261373
Test FAILed.
Refer to this link for build results (access rights to CI server needed):
Github user SparkQA commented on the pull request:
https://github.com/apache/spark/pull/4173#issuecomment-71261364
[Test build #26034 has
finished](https://amplab.cs.berkeley.edu/jenkins/job/SparkPullRequestBuilder/26034/consoleFull)
for PR 4173 at commit
Github user vanzin commented on a diff in the pull request:
https://github.com/apache/spark/pull/4155#discussion_r23473536
--- Diff:
core/src/test/scala/org/apache/spark/scheduler/DAGSchedulerSuite.scala ---
@@ -19,12 +19,13 @@ package org.apache.spark.scheduler
import
Github user vanzin commented on a diff in the pull request:
https://github.com/apache/spark/pull/4155#discussion_r23474074
--- Diff:
core/src/main/scala/org/apache/spark/scheduler/OutputCommitCoordinator.scala ---
@@ -0,0 +1,178 @@
+/*
+ * Licensed to the Apache Software
Github user mccheah commented on the pull request:
https://github.com/apache/spark/pull/4155#issuecomment-71253931
I'm also concerned about the performance ramifications of this. We need to
run performance benchmarks. However, the only critical path that is affected by
this are tasks
Github user AmplabJenkins commented on the pull request:
https://github.com/apache/spark/pull/4139#issuecomment-71261303
Test FAILed.
Refer to this link for build results (access rights to CI server needed):
Github user SparkQA commented on the pull request:
https://github.com/apache/spark/pull/4139#issuecomment-71261296
[Test build #26035 has
finished](https://amplab.cs.berkeley.edu/jenkins/job/SparkPullRequestBuilder/26035/consoleFull)
for PR 4139 at commit
Github user vanzin commented on a diff in the pull request:
https://github.com/apache/spark/pull/3884#discussion_r23470546
--- Diff: python/pyspark/rdd.py ---
@@ -141,6 +141,17 @@ def id(self):
def __repr__(self):
return self._jrdd.toString()
+
Github user SparkQA commented on the pull request:
https://github.com/apache/spark/pull/4173#issuecomment-71244959
[Test build #26033 has
started](https://amplab.cs.berkeley.edu/jenkins/job/SparkPullRequestBuilder/26033/consoleFull)
for PR 4173 at commit
Github user AmplabJenkins commented on the pull request:
https://github.com/apache/spark/pull/4173#issuecomment-71246679
Test FAILed.
Refer to this link for build results (access rights to CI server needed):
Github user SparkQA commented on the pull request:
https://github.com/apache/spark/pull/4173#issuecomment-71246730
[Test build #26034 has
started](https://amplab.cs.berkeley.edu/jenkins/job/SparkPullRequestBuilder/26034/consoleFull)
for PR 4173 at commit
Github user AmplabJenkins commented on the pull request:
https://github.com/apache/spark/pull/3916#issuecomment-71255189
Test PASSed.
Refer to this link for build results (access rights to CI server needed):
Github user ankurdave commented on the pull request:
https://github.com/apache/spark/pull/4136#issuecomment-71265542
@JoshRosen No, it doesn't seem to trigger the Snappy error! After the
previous attempted fix (#1763, 9b225ac3072de522b40b46aba6df1f1c231f13ef), the
GraphX unit tests
Github user davies commented on a diff in the pull request:
https://github.com/apache/spark/pull/4173#discussion_r23488297
--- Diff: sql/core/src/main/scala/org/apache/spark/sql/DataFrame.scala ---
@@ -0,0 +1,273 @@
+/*
+* Licensed to the Apache Software Foundation (ASF)
Github user mengxr commented on the pull request:
https://github.com/apache/spark/pull/4183#issuecomment-71276246
I agree that vectors must have the same length and we should check it. It
may not be necessary to change the implementation. I saw couple performance
issues in your code.
Github user dbtsai commented on a diff in the pull request:
https://github.com/apache/spark/pull/4140#discussion_r23486231
--- Diff:
mllib/src/main/scala/org/apache/spark/mllib/feature/StandardScaler.scala ---
@@ -61,20 +61,30 @@ class StandardScaler(withMean: Boolean, withStd:
Github user dbtsai commented on the pull request:
https://github.com/apache/spark/pull/4140#issuecomment-71281849
For the unit-test part, is it possible not to change too much? Also, it
will be easier to debug if the assertion is in the test instead of abstract
out. For example,
Github user davies commented on a diff in the pull request:
https://github.com/apache/spark/pull/4173#discussion_r23488167
--- Diff: sql/core/src/main/scala/org/apache/spark/sql/DataFrame.scala ---
@@ -0,0 +1,273 @@
+/*
+* Licensed to the Apache Software Foundation (ASF)
Github user davies commented on a diff in the pull request:
https://github.com/apache/spark/pull/4173#discussion_r23488501
--- Diff: sql/core/src/main/scala/org/apache/spark/sql/DataFrame.scala ---
@@ -0,0 +1,273 @@
+/*
+* Licensed to the Apache Software Foundation (ASF)
Github user vanzin commented on the pull request:
https://github.com/apache/spark/pull/4082#issuecomment-71272446
LGTM
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled
Github user SparkQA commented on the pull request:
https://github.com/apache/spark/pull/1290#issuecomment-71284930
[Test build #26036 has
started](https://amplab.cs.berkeley.edu/jenkins/job/SparkPullRequestBuilder/26036/consoleFull)
for PR 1290 at commit
Github user davies commented on a diff in the pull request:
https://github.com/apache/spark/pull/4173#discussion_r23488078
--- Diff: sql/core/src/main/scala/org/apache/spark/sql/DataFrame.scala ---
@@ -0,0 +1,273 @@
+/*
+* Licensed to the Apache Software Foundation (ASF)
Github user dbtsai commented on a diff in the pull request:
https://github.com/apache/spark/pull/4140#discussion_r23485163
--- Diff:
mllib/src/main/scala/org/apache/spark/mllib/feature/StandardScaler.scala ---
@@ -61,20 +61,30 @@ class StandardScaler(withMean: Boolean, withStd:
Github user ksakellis commented on the pull request:
https://github.com/apache/spark/pull/4082#issuecomment-71282577
LGTM - nice addition.
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have
GitHub user srowen opened a pull request:
https://github.com/apache/spark/pull/4181
SPARK-984 [BUILD] SPARK_TOOLS_JAR not set if multiple tools jars exists
Given the discussion in https://issues.apache.org/jira/browse/SPARK-984,
this seems to be the outcome, but I'm not 100% sure
Github user SparkQA commented on the pull request:
https://github.com/apache/spark/pull/3519#issuecomment-71189621
[Test build #26025 has
started](https://amplab.cs.berkeley.edu/jenkins/job/SparkPullRequestBuilder/26025/consoleFull)
for PR 3519 at commit
Github user AmplabJenkins commented on the pull request:
https://github.com/apache/spark/pull/4181#issuecomment-71194064
Test FAILed.
Refer to this link for build results (access rights to CI server needed):
Github user SparkQA commented on the pull request:
https://github.com/apache/spark/pull/4181#issuecomment-71194057
[Test build #26024 has
finished](https://amplab.cs.berkeley.edu/jenkins/job/SparkPullRequestBuilder/26024/consoleFull)
for PR 4181 at commit
Github user srowen closed the pull request at:
https://github.com/apache/spark/pull/4181
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is
Github user srowen commented on the pull request:
https://github.com/apache/spark/pull/4181#issuecomment-71194299
Ah. This makes Mima stop working. OK, this isn't an option!
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as
GitHub user rxin opened a pull request:
https://github.com/apache/spark/pull/4173
[SPARK-5097][WIP] DataFrame as the common abstraction for structured data
This is early work in progress. I am submitting the PR mainly wanted to get
Jenkins to run through the tests so I don't have
Github user SparkQA commented on the pull request:
https://github.com/apache/spark/pull/4136#issuecomment-71160721
[Test build #26004 has
finished](https://amplab.cs.berkeley.edu/jenkins/job/SparkPullRequestBuilder/26004/consoleFull)
for PR 4136 at commit
GitHub user zsxwing opened a pull request:
https://github.com/apache/spark/pull/4174
[SPARK-5214][Test] Add a test to demonstrate EventLoop can be stopped in
the event loop thread
You can merge this pull request into a Git repository by running:
$ git pull
Github user AmplabJenkins commented on the pull request:
https://github.com/apache/spark/pull/3247#issuecomment-71162789
Test FAILed.
Refer to this link for build results (access rights to CI server needed):
Github user SparkQA commented on the pull request:
https://github.com/apache/spark/pull/3247#issuecomment-71162787
[Test build #26013 has
finished](https://amplab.cs.berkeley.edu/jenkins/job/SparkPullRequestBuilder/26013/consoleFull)
for PR 3247 at commit
Github user SparkQA commented on the pull request:
https://github.com/apache/spark/pull/3247#issuecomment-71162714
[Test build #26013 has
started](https://amplab.cs.berkeley.edu/jenkins/job/SparkPullRequestBuilder/26013/consoleFull)
for PR 3247 at commit
Github user SparkQA commented on the pull request:
https://github.com/apache/spark/pull/4170#issuecomment-71166652
[Test build #26015 has
started](https://amplab.cs.berkeley.edu/jenkins/job/SparkPullRequestBuilder/26015/consoleFull)
for PR 4170 at commit
GitHub user Leolh opened a pull request:
https://github.com/apache/spark/pull/4176
[SPARK-3650][GraphX] There will be an ArrayIndexOutOfBoundsException if ...
...the format of the source file is wrong
There will be an ArrayIndexOutOfBoundsException if the format of the
Github user OopsOutOfMemory commented on the pull request:
https://github.com/apache/spark/pull/4175#issuecomment-71167869
/cc @scwf @chenghao-intel
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does
Github user JoshRosen commented on the pull request:
https://github.com/apache/spark/pull/4179#issuecomment-71294854
@andrewor14 since you reviewed the other PR for `SPARK_CONF_DIR`, can you
take a quick look at this and #4177 to see if we want to pull it in for 1.2.1?
---
If your
Github user ankurdave commented on the pull request:
https://github.com/apache/spark/pull/4136#issuecomment-71297997
Oh, thanks! Looks like that was the problem all along; stopping the
SparkContext fixes the problem. I'm going to merge this with the amended test
now.
---
If your
Github user ankurdave commented on the pull request:
https://github.com/apache/spark/pull/4136#issuecomment-71286155
@JoshRosen Actually, it seems the test failures still occur, but only when
I add a [unit
Github user shivaram commented on the pull request:
https://github.com/apache/spark/pull/4038#issuecomment-71286989
@voukka @nchammas - This high level goal looks fine to me. However I the
function get_hostname is being called on all instances (its inside a loop) in
many cases. I
Github user shivaram commented on the pull request:
https://github.com/apache/spark/pull/3986#issuecomment-71289144
@nchammas @GenTang - The `logging.basicConfig` seems to have been around
since the very beginning [1]. I don't know much about Python so I can't
recommend keeping it or
Github user SparkQA commented on the pull request:
https://github.com/apache/spark/pull/1290#issuecomment-71290964
[Test build #26036 has
finished](https://amplab.cs.berkeley.edu/jenkins/job/SparkPullRequestBuilder/26036/consoleFull)
for PR 1290 at commit
Github user AmplabJenkins commented on the pull request:
https://github.com/apache/spark/pull/1290#issuecomment-71290975
Test PASSed.
Refer to this link for build results (access rights to CI server needed):
Github user JoshRosen commented on the pull request:
https://github.com/apache/spark/pull/3884#issuecomment-71292584
@vanzin Thanks for looking this over. The Python `RDD` objects themselves
are never actually serialized and are used internally in a way that's slightly
different
Github user nchammas commented on the pull request:
https://github.com/apache/spark/pull/3884#issuecomment-71298244
Thank you @JoshRosen for working on usability issues like this.
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub
Github user ilganeli closed the pull request at:
https://github.com/apache/spark/pull/3518
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is
Github user ilganeli commented on the pull request:
https://github.com/apache/spark/pull/3518#issuecomment-71299062
Hey @pwendell - not a problem. The solutions are similar but Reynold's has
fewer moving parts. I appreciate the recognition.
---
If your project is set up for it, you
Github user JoshRosen commented on the pull request:
https://github.com/apache/spark/pull/4136#issuecomment-71293352
@ankurdave The exception from the new unit test sounds suspiciously similar
to https://issues.apache.org/jira/browse/SPARK-4133. Your new test creates a
new `sc`
Github user SparkQA commented on the pull request:
https://github.com/apache/spark/pull/3884#issuecomment-71294206
[Test build #26037 has
started](https://amplab.cs.berkeley.edu/jenkins/job/SparkPullRequestBuilder/26037/consoleFull)
for PR 3884 at commit
Github user AmplabJenkins commented on the pull request:
https://github.com/apache/spark/pull/3884#issuecomment-71297060
Test PASSed.
Refer to this link for build results (access rights to CI server needed):
Github user SparkQA commented on the pull request:
https://github.com/apache/spark/pull/3884#issuecomment-71297055
[Test build #26037 has
finished](https://amplab.cs.berkeley.edu/jenkins/job/SparkPullRequestBuilder/26037/consoleFull)
for PR 3884 at commit
Github user asfgit closed the pull request at:
https://github.com/apache/spark/pull/4136
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is
Github user asfgit closed the pull request at:
https://github.com/apache/spark/pull/3884
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is
Github user JoshRosen commented on the pull request:
https://github.com/apache/spark/pull/3884#issuecomment-71295041
Alright, I've merged this into `master` (1.3.0) and `branch-1.2` (1.2.1).
---
If your project is set up for it, you can reply to this email and have your
reply appear
Github user SparkQA commented on the pull request:
https://github.com/apache/spark/pull/4140#issuecomment-71297633
[Test build #26038 has
started](https://amplab.cs.berkeley.edu/jenkins/job/SparkPullRequestBuilder/26038/consoleFull)
for PR 4140 at commit
Github user ogeagla commented on the pull request:
https://github.com/apache/spark/pull/4140#issuecomment-71297662
@dbtsai that makes sense. I've changed this back in latest commit.
---
If your project is set up for it, you can reply to this email and have your
reply appear on
Github user ankurdave commented on the pull request:
https://github.com/apache/spark/pull/4136#issuecomment-71298562
Merged into master branch-1.2.
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does
Github user SparkQA commented on the pull request:
https://github.com/apache/spark/pull/4173#issuecomment-71159715
[Test build #26008 has
started](https://amplab.cs.berkeley.edu/jenkins/job/SparkPullRequestBuilder/26008/consoleFull)
for PR 4173 at commit
Github user rxin commented on the pull request:
https://github.com/apache/spark/pull/4134#issuecomment-71159737
Thanks for doing it. I took a quick look at this. While it does reduce the
LOC, I feel the change is not necessary and actually makes the code harder to
understand with the
Github user SparkQA commented on the pull request:
https://github.com/apache/spark/pull/4174#issuecomment-71161124
[Test build #26009 has
started](https://amplab.cs.berkeley.edu/jenkins/job/SparkPullRequestBuilder/26009/consoleFull)
for PR 4174 at commit
Github user SparkQA commented on the pull request:
https://github.com/apache/spark/pull/4173#issuecomment-71161215
[Test build #26010 has
finished](https://amplab.cs.berkeley.edu/jenkins/job/SparkPullRequestBuilder/26010/consoleFull)
for PR 4173 at commit
Github user AmplabJenkins commented on the pull request:
https://github.com/apache/spark/pull/4173#issuecomment-71161217
Test FAILed.
Refer to this link for build results (access rights to CI server needed):
Github user AmplabJenkins commented on the pull request:
https://github.com/apache/spark/pull/4173#issuecomment-71162544
Test FAILed.
Refer to this link for build results (access rights to CI server needed):
Github user cloud-fan commented on a diff in the pull request:
https://github.com/apache/spark/pull/4055#discussion_r23437961
--- Diff: core/src/main/scala/org/apache/spark/scheduler/Task.scala ---
@@ -106,7 +106,22 @@ private[spark] abstract class Task[T](val stageId:
Int, var
GitHub user OopsOutOfMemory opened a pull request:
https://github.com/apache/spark/pull/4175
[SPARK-3298][SQL] Add flag control overwrite registerAsTable /
registerTempTable
https://issues.apache.org/jira/browse/SPARK-3298
add a flag `allowOverwrite` to control
Github user adrian-wang commented on the pull request:
https://github.com/apache/spark/pull/4057#issuecomment-71167347
Yes, I moved my work to FunctionArgumentConversion, and since #4040 is
reverted due to conflicts, I added the code together here. So I leave
Coalesce() untouched,
Github user SparkQA commented on the pull request:
https://github.com/apache/spark/pull/4178#issuecomment-71181359
[Test build #26020 has
finished](https://amplab.cs.berkeley.edu/jenkins/job/SparkPullRequestBuilder/26020/consoleFull)
for PR 4178 at commit
Github user viirya commented on the pull request:
https://github.com/apache/spark/pull/4158#issuecomment-71181584
@chenghao-intel overall it looks good for me except for small comments.
---
If your project is set up for it, you can reply to this email and have your
reply appear on
1 - 100 of 210 matches
Mail list logo