Github user davies commented on the issue:
https://github.com/apache/spark/pull/13701
Merging this into master and 2.0, thanks!
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
Github user davies commented on a diff in the pull request:
https://github.com/apache/spark/pull/14500#discussion_r74132414
--- Diff:
sql/core/src/main/scala/org/apache/spark/sql/execution/command/ddl.scala ---
@@ -425,6 +430,111 @@ case class AlterTableDropPartitionCommand
Github user davies commented on a diff in the pull request:
https://github.com/apache/spark/pull/14500#discussion_r74132132
--- Diff:
sql/core/src/main/scala/org/apache/spark/sql/execution/command/ddl.scala ---
@@ -425,6 +430,111 @@ case class AlterTableDropPartitionCommand
Github user davies commented on the issue:
https://github.com/apache/spark/pull/13701
LGTM, could you fix the conflict (should be trivial)?
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have
Github user davies commented on the issue:
https://github.com/apache/spark/pull/14500
Merging into master, thanks!
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled
Github user davies commented on a diff in the pull request:
https://github.com/apache/spark/pull/14500#discussion_r74100170
--- Diff:
sql/core/src/main/scala/org/apache/spark/sql/execution/command/ddl.scala ---
@@ -425,6 +430,110 @@ case class AlterTableDropPartitionCommand
Github user davies commented on a diff in the pull request:
https://github.com/apache/spark/pull/14500#discussion_r74099592
--- Diff:
sql/core/src/main/scala/org/apache/spark/sql/execution/command/ddl.scala ---
@@ -425,6 +430,111 @@ case class AlterTableDropPartitionCommand
Github user davies commented on the issue:
https://github.com/apache/spark/pull/14540
LGTM, merging this into master, thanks!
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
Github user davies commented on a diff in the pull request:
https://github.com/apache/spark/pull/14500#discussion_r74094542
--- Diff:
sql/core/src/main/scala/org/apache/spark/sql/execution/command/ddl.scala ---
@@ -425,6 +430,110 @@ case class AlterTableDropPartitionCommand
Github user davies commented on a diff in the pull request:
https://github.com/apache/spark/pull/14500#discussion_r74094235
--- Diff:
sql/core/src/main/scala/org/apache/spark/sql/execution/command/ddl.scala ---
@@ -425,6 +430,111 @@ case class AlterTableDropPartitionCommand
GitHub user davies opened a pull request:
https://github.com/apache/spark/pull/14548
[SPARK-16958] [SQL] Reuse subqueries within the same query
## What changes were proposed in this pull request?
There could be multiple subqueries that generate same results, we could
re
Github user davies closed the pull request at:
https://github.com/apache/spark/pull/14545
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature
Github user davies commented on a diff in the pull request:
https://github.com/apache/spark/pull/14469#discussion_r73960965
--- Diff: python/pyspark/sql/session.py ---
@@ -384,17 +384,15 @@ def _createFromLocal(self, data, schema):
if schema is None or isinstance
GitHub user davies opened a pull request:
https://github.com/apache/spark/pull/14545
[SPARK-11150] [SQL] Dynamic Partition Pruning
## What changes were proposed in this pull request?
This PR introduces a new feature for Spark SQL: dynamic partition pruning,
which could
Github user davies commented on a diff in the pull request:
https://github.com/apache/spark/pull/13680#discussion_r73951262
--- Diff:
sql/catalyst/src/test/scala/org/apache/spark/sql/catalyst/util/UnsafeArraySuite.scala
---
@@ -18,27 +18,131 @@
package
Github user davies commented on a diff in the pull request:
https://github.com/apache/spark/pull/14500#discussion_r73945744
--- Diff:
sql/core/src/main/scala/org/apache/spark/sql/execution/command/ddl.scala ---
@@ -425,6 +430,110 @@ case class AlterTableDropPartitionCommand
Github user davies commented on a diff in the pull request:
https://github.com/apache/spark/pull/14500#discussion_r73945656
--- Diff:
sql/core/src/main/scala/org/apache/spark/sql/execution/command/ddl.scala ---
@@ -425,6 +430,110 @@ case class AlterTableDropPartitionCommand
Github user davies commented on a diff in the pull request:
https://github.com/apache/spark/pull/13701#discussion_r73934292
--- Diff:
sql/core/src/main/scala/org/apache/spark/sql/execution/datasources/parquet/ParquetFileFormat.scala
---
@@ -357,10 +358,27 @@ private[sql] class
Github user davies commented on a diff in the pull request:
https://github.com/apache/spark/pull/13701#discussion_r73934243
--- Diff:
sql/core/src/main/scala/org/apache/spark/sql/execution/DataSourceScanExec.scala
---
@@ -213,7 +213,9 @@ private[sql] case class FileSourceScanExec
Github user davies commented on a diff in the pull request:
https://github.com/apache/spark/pull/14469#discussion_r73932092
--- Diff: python/pyspark/sql/session.py ---
@@ -432,14 +430,9 @@ def createDataFrame(self, data, schema=None,
samplingRatio=None):
``byte
Github user davies commented on the issue:
https://github.com/apache/spark/pull/14454
LGTM
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so
Github user davies commented on a diff in the pull request:
https://github.com/apache/spark/pull/14266#discussion_r73928683
--- Diff:
sql/core/src/test/scala/org/apache/spark/sql/execution/benchmark/AggregateBenchmark.scala
---
@@ -1078,6 +1078,146 @@ class AggregateBenchmark
Github user davies commented on a diff in the pull request:
https://github.com/apache/spark/pull/13680#discussion_r73926983
--- Diff:
sql/catalyst/src/main/java/org/apache/spark/sql/catalyst/expressions/UnsafeArrayData.java
---
@@ -25,55 +25,57 @@
import
Github user davies commented on the issue:
https://github.com/apache/spark/pull/14513
It's not necessary, but more clear (consistent)
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have
Github user davies commented on the issue:
https://github.com/apache/spark/pull/14513
We should also update VectorizedColumnReader.decodeDictionaryIds() to use
the new method.
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub
Github user davies commented on the issue:
https://github.com/apache/spark/pull/14176
Let's hold on this, if we are going to have single implementation for fast
hash map (based on the benchmark result in another PR), do need to merge this
fancy implementation choosing. cc @rxin
Github user davies commented on a diff in the pull request:
https://github.com/apache/spark/pull/13701#discussion_r73756701
--- Diff:
sql/core/src/main/scala/org/apache/spark/sql/execution/DataSourceScanExec.scala
---
@@ -199,6 +209,19 @@ private[sql] case class
Github user davies commented on a diff in the pull request:
https://github.com/apache/spark/pull/14500#discussion_r73753834
--- Diff:
sql/core/src/test/scala/org/apache/spark/sql/execution/command/DDLSuite.scala
---
@@ -827,6 +827,45 @@ class DDLSuite extends QueryTest
Github user davies commented on a diff in the pull request:
https://github.com/apache/spark/pull/13680#discussion_r73753275
--- Diff:
sql/catalyst/src/main/java/org/apache/spark/sql/catalyst/expressions/UnsafeArrayData.java
---
@@ -25,30 +25,36 @@
import
Github user davies commented on a diff in the pull request:
https://github.com/apache/spark/pull/14500#discussion_r73729864
--- Diff:
sql/core/src/main/scala/org/apache/spark/sql/execution/SparkSqlParser.scala ---
@@ -409,6 +409,18 @@ class SparkSqlAstBuilder(conf: SQLConf
Github user davies commented on a diff in the pull request:
https://github.com/apache/spark/pull/13701#discussion_r73728732
--- Diff:
sql/core/src/test/scala/org/apache/spark/sql/execution/datasources/parquet/ParquetFilterSuite.scala
---
@@ -368,73 +378,75 @@ class
Github user davies commented on a diff in the pull request:
https://github.com/apache/spark/pull/13680#discussion_r73727756
--- Diff:
sql/catalyst/src/main/java/org/apache/spark/sql/catalyst/expressions/UnsafeArrayData.java
---
@@ -25,30 +25,36 @@
import
Github user davies commented on a diff in the pull request:
https://github.com/apache/spark/pull/13680#discussion_r73727463
--- Diff:
sql/catalyst/src/main/java/org/apache/spark/sql/catalyst/expressions/UnsafeArrayData.java
---
@@ -25,30 +25,36 @@
import
Github user davies commented on the issue:
https://github.com/apache/spark/pull/14500
@yhuai Just checked the repair.q, it's kind of useless, already covered by
out unit test, we could just ignore it.
---
If your project is set up for it, you can reply to this email and have your
Github user davies commented on a diff in the pull request:
https://github.com/apache/spark/pull/14266#discussion_r73616369
--- Diff:
sql/core/src/test/scala/org/apache/spark/sql/execution/benchmark/AggregateBenchmark.scala
---
@@ -576,4 +576,605 @@ class AggregateBenchmark
GitHub user davies opened a pull request:
https://github.com/apache/spark/pull/14500
[SPARK-] SQL DDL: MSCK REPAIR TABLE
## What changes were proposed in this pull request?
MSCK REPAIR TABLE could be used to recover the partitions in external
catalog based on partitions
Github user davies commented on the issue:
https://github.com/apache/spark/pull/14500
@yhuai Could you help to generate the golden result for this suite?
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project
Github user davies commented on a diff in the pull request:
https://github.com/apache/spark/pull/13680#discussion_r73594310
--- Diff:
sql/catalyst/src/main/java/org/apache/spark/sql/catalyst/expressions/UnsafeArrayData.java
---
@@ -25,30 +25,36 @@
import
Github user davies commented on the issue:
https://github.com/apache/spark/pull/14487
LGTM, merging into master, thanks!
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled
Github user davies commented on a diff in the pull request:
https://github.com/apache/spark/pull/13680#discussion_r73569234
--- Diff:
sql/catalyst/src/main/java/org/apache/spark/sql/catalyst/expressions/UnsafeArrayData.java
---
@@ -25,30 +25,36 @@
import
Github user davies commented on a diff in the pull request:
https://github.com/apache/spark/pull/13680#discussion_r73569132
--- Diff:
sql/catalyst/src/main/java/org/apache/spark/sql/catalyst/expressions/UnsafeArrayData.java
---
@@ -25,30 +25,36 @@
import
Github user davies commented on a diff in the pull request:
https://github.com/apache/spark/pull/13680#discussion_r73568470
--- Diff:
sql/catalyst/src/main/java/org/apache/spark/sql/catalyst/expressions/UnsafeArrayData.java
---
@@ -25,30 +25,36 @@
import
Github user davies commented on a diff in the pull request:
https://github.com/apache/spark/pull/13680#discussion_r73567644
--- Diff:
sql/catalyst/src/main/java/org/apache/spark/sql/catalyst/expressions/UnsafeArrayData.java
---
@@ -59,21 +65,16 @@
// The 4-bytes header
Github user davies commented on a diff in the pull request:
https://github.com/apache/spark/pull/13680#discussion_r73567215
--- Diff:
sql/catalyst/src/main/java/org/apache/spark/sql/catalyst/expressions/UnsafeArrayData.java
---
@@ -25,30 +25,36 @@
import
Github user davies commented on a diff in the pull request:
https://github.com/apache/spark/pull/13680#discussion_r73566559
--- Diff:
sql/catalyst/src/main/java/org/apache/spark/sql/catalyst/expressions/UnsafeArrayData.java
---
@@ -25,30 +25,36 @@
import
Github user davies commented on a diff in the pull request:
https://github.com/apache/spark/pull/13680#discussion_r73565969
--- Diff:
sql/catalyst/src/main/java/org/apache/spark/sql/catalyst/expressions/UnsafeArrayData.java
---
@@ -25,30 +25,36 @@
import
Github user davies commented on a diff in the pull request:
https://github.com/apache/spark/pull/13680#discussion_r73565692
--- Diff:
sql/catalyst/src/test/scala/org/apache/spark/sql/catalyst/util/UnsafeArraySuite.scala
---
@@ -18,27 +18,131 @@
package
Github user davies commented on a diff in the pull request:
https://github.com/apache/spark/pull/13680#discussion_r73565092
--- Diff:
sql/core/src/test/scala/org/apache/spark/sql/execution/columnar/ColumnTypeSuite.scala
---
@@ -73,8 +73,8 @@ class ColumnTypeSuite extends
Github user davies commented on the issue:
https://github.com/apache/spark/pull/14241
Merging this into master, thanks!
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled
Github user davies commented on the issue:
https://github.com/apache/spark/pull/12983
If the number of iteration is not huge, it does not matter to use range()
or xrange() in Python 2 (especially when you use it together with `for`).
That's said I'm not a fan of this change
Github user davies commented on a diff in the pull request:
https://github.com/apache/spark/pull/13701#discussion_r73383148
--- Diff:
sql/core/src/test/scala/org/apache/spark/sql/execution/datasources/parquet/ParquetFilterSuite.scala
---
@@ -527,4 +538,54 @@ class
Github user davies commented on the issue:
https://github.com/apache/spark/pull/14241
@hvanhovell Have you finished your round of review?
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have
GitHub user davies opened a pull request:
https://github.com/apache/spark/pull/14469
[SPARK-16700] [PYSPARK] [SQL] create DataFrame from dict/Row with schema
## What changes were proposed in this pull request?
In 2.0, we verify the data type against schema for every row
Github user davies commented on a diff in the pull request:
https://github.com/apache/spark/pull/13701#discussion_r73258703
--- Diff:
sql/core/src/main/scala/org/apache/spark/sql/execution/datasources/parquet/ParquetFileFormat.scala
---
@@ -581,62 +586,6 @@ private[sql] object
Github user davies commented on the issue:
https://github.com/apache/spark/pull/14241
LGTM
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so
Github user davies commented on a diff in the pull request:
https://github.com/apache/spark/pull/13701#discussion_r73247832
--- Diff:
sql/core/src/main/scala/org/apache/spark/sql/execution/datasources/parquet/ParquetFileFormat.scala
---
@@ -581,62 +586,6 @@ private[sql] object
Github user davies commented on a diff in the pull request:
https://github.com/apache/spark/pull/13701#discussion_r73247716
--- Diff:
sql/core/src/main/scala/org/apache/spark/sql/execution/datasources/FileSourceStrategy.scala
---
@@ -85,8 +85,15 @@ private[sql] object
Github user davies commented on the issue:
https://github.com/apache/spark/pull/13701
@gatorsmile In order to merge this patch sooner, it's better to only have
related changes to fix the regression. We can clean the dead code later.
---
If your project is set up for it, you can
Github user davies commented on a diff in the pull request:
https://github.com/apache/spark/pull/14241#discussion_r73231816
--- Diff:
sql/core/src/main/scala/org/apache/spark/sql/execution/ExistingRDD.scala ---
@@ -275,62 +272,161 @@ private[sql] case class RowDataSourceScanExec
Github user davies commented on a diff in the pull request:
https://github.com/apache/spark/pull/14464#discussion_r73231143
--- Diff:
sql/core/src/main/scala/org/apache/spark/sql/execution/joins/HashedRelation.scala
---
@@ -459,8 +459,8 @@ private[execution] final class
Github user davies commented on a diff in the pull request:
https://github.com/apache/spark/pull/14241#discussion_r73228065
--- Diff:
sql/core/src/main/scala/org/apache/spark/sql/execution/ExistingRDD.scala ---
@@ -275,62 +272,161 @@ private[sql] case class RowDataSourceScanExec
Github user davies commented on the issue:
https://github.com/apache/spark/pull/14465
@maver1ck Thanks for sending out this, I'd prefer to merge #13701, there
are already lots of discussions there.
---
If your project is set up for it, you can reply to this email and have your
reply
Github user davies commented on a diff in the pull request:
https://github.com/apache/spark/pull/14241#discussion_r73226713
--- Diff:
sql/core/src/main/scala/org/apache/spark/sql/execution/ExistingRDD.scala ---
@@ -275,62 +272,161 @@ private[sql] case class RowDataSourceScanExec
GitHub user davies opened a pull request:
https://github.com/apache/spark/pull/14464
[SPARK-16802] [SQL] fix overflow in LongToUnsafeRowMap
## What changes were proposed in this pull request?
This patch fix the overflow in LongToUnsafeRowMap when the range of key is
very
Github user davies commented on a diff in the pull request:
https://github.com/apache/spark/pull/13701#discussion_r73207167
--- Diff:
sql/core/src/main/scala/org/apache/spark/sql/execution/datasources/FileSourceStrategy.scala
---
@@ -85,8 +85,15 @@ private[sql] object
Github user davies commented on the issue:
https://github.com/apache/spark/pull/13701
@viirya In order to have a unit test for this (otherwise it will be broken
again in future), we could add some counter in vectorwized parquet reader for
row groups for test purpose, then use
Github user davies commented on a diff in the pull request:
https://github.com/apache/spark/pull/13701#discussion_r73206857
--- Diff:
sql/core/src/main/scala/org/apache/spark/sql/execution/datasources/parquet/ParquetFileFormat.scala
---
@@ -357,6 +357,11 @@ private[sql] class
Github user davies commented on the issue:
https://github.com/apache/spark/pull/14446
The changes looks good to me. Could you post the numbers of benchmark in PR
description ?
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub
Github user davies commented on the issue:
https://github.com/apache/spark/pull/13778
LGTM, merging this into master and 2.0 branch, thanks!
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have
Github user davies commented on the issue:
https://github.com/apache/spark/pull/14442
LGTM
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so
Github user davies commented on the issue:
https://github.com/apache/spark/pull/13107
LGTM,
Merging this into master and 2.0, thanks!
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have
Github user davies commented on the issue:
https://github.com/apache/spark/pull/13977
LGTM,
Merging this into master and 2.0, thanks!
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have
Github user davies commented on the issue:
https://github.com/apache/spark/pull/13963
Merging this into master and 2.0, thanks!
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
Github user davies commented on the issue:
https://github.com/apache/spark/pull/13878
Merging this into master and 2.0, thanks!
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
GitHub user davies opened a pull request:
https://github.com/apache/spark/pull/13948
[SPARK-16259] [PYSPARK] cleanup options in DataFrame read/write API
## What changes were proposed in this pull request?
There are some duplicated code for options in DataFrame reader/writer
Github user davies commented on the issue:
https://github.com/apache/spark/pull/13948
cc @rxin
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so
Github user davies commented on the issue:
https://github.com/apache/spark/pull/13931
LGTM,
Merging this into master and 2.0, thanks!
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have
Github user davies commented on the issue:
https://github.com/apache/spark/pull/13027
@tgravescs For minor releases, we (the community) usually do not put much
effort on the QA, it's risky to pull in large change (like this one) it. At
least, I don't have enough confidence to merge
Github user davies commented on a diff in the pull request:
https://github.com/apache/spark/pull/13107#discussion_r68620470
--- Diff:
core/src/main/java/org/apache/spark/shuffle/sort/ShuffleExternalSorter.java ---
@@ -72,7 +72,10 @@
private final TaskContext taskContext
Github user davies commented on the issue:
https://github.com/apache/spark/pull/13900
@dongjoon-hyun had also merged into 1.5 branch
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have
Github user davies commented on the issue:
https://github.com/apache/spark/pull/13887
Merged into master, thanks
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled
Github user davies commented on the issue:
https://github.com/apache/spark/pull/13902
Merged into 1.6
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so
Github user davies commented on the issue:
https://github.com/apache/spark/pull/13900
@dongjoon-hyun Could you send a patch for 1.6?
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have
Github user davies commented on the issue:
https://github.com/apache/spark/pull/13887
Reverted, will merge this again once it passed jenkins.
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have
Github user davies commented on the issue:
https://github.com/apache/spark/pull/13887
Sorry, the jenkins has not finished ...
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
Github user davies commented on the issue:
https://github.com/apache/spark/pull/13887
LGTM, merging this into master.
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled
Github user davies commented on a diff in the pull request:
https://github.com/apache/spark/pull/13887#discussion_r68474760
--- Diff:
sql/core/src/main/scala/org/apache/spark/sql/execution/columnar/InMemoryTableScanExec.scala
---
@@ -79,6 +79,11 @@ private[sql] case class
Github user davies commented on the issue:
https://github.com/apache/spark/pull/13887
Let's go with current patch, I will review it now. Those things could be
considered later.
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub
Github user davies commented on the issue:
https://github.com/apache/spark/pull/13887
@dongjoon-hyun That's a good point, the current patch is better for
performance actually
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub
Github user davies commented on the issue:
https://github.com/apache/spark/pull/13887
@dongjoon-hyun Yes, 2) should check the constraints to make it idempotent
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your
Github user davies commented on the issue:
https://github.com/apache/spark/pull/13887
@dongjoon-hyun There is only single predicate in Filter, it could be AND or
OR, so it means we could control the order. For this case, I'm not sure the
inserted GreaterThanOrEqual/LessThanOrEqual
Github user davies commented on the issue:
https://github.com/apache/spark/pull/13900
LGTM
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so
Github user davies commented on the issue:
https://github.com/apache/spark/pull/13887
For any IN that have more than one expression, we could add another
GreaterThanOrEqual/LessThanOrEqual (not replace the IN).
For 2, it's not that obvious yet, we can do that later
Github user davies commented on a diff in the pull request:
https://github.com/apache/spark/pull/13900#discussion_r68468749
--- Diff: sql/core/src/main/scala/org/apache/spark/sql/Dataset.scala ---
@@ -1908,7 +1908,7 @@ class Dataset[T] private[sql](
// All columns
Github user davies commented on the issue:
https://github.com/apache/spark/pull/13883
@rxin Could you review this?
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled
Github user davies commented on the issue:
https://github.com/apache/spark/pull/13788
Merged into 1.6, 2.0 and master
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled
Github user davies commented on the issue:
https://github.com/apache/spark/pull/13887
BTW, we could use constraints to implement this.
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have
Github user davies commented on the issue:
https://github.com/apache/spark/pull/13887
@dongjoon-hyun Thanks for the patch, this optimization sounds reasonable.
I'm thinking of is it possible to make the optimization for IN/INSET more
general. We could have a optimizer
Github user davies commented on the issue:
https://github.com/apache/spark/pull/13883
Jenkins, retest this please
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled
Github user davies commented on the issue:
https://github.com/apache/spark/pull/13883
https://gist.github.com/vlad17/964c0a93510d79cb130c33700f6139b7
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does
301 - 400 of 6045 matches
Mail list logo