Github user rxin commented on the issue:
https://github.com/apache/spark/pull/16875
@bogdanrdc can you close this? It won't auto close because it is not merged
in master.
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as
Github user rxin commented on the issue:
https://github.com/apache/spark/pull/16875
Merging in branch-2.1.
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so
Github user rxin commented on a diff in the pull request:
https://github.com/apache/spark/pull/16864#discussion_r100503141
--- Diff:
common/sketch/src/main/java/org/apache/spark/util/sketch/BloomFilter.java ---
@@ -81,6 +81,11 @@ int getVersionNumber() {
public abstract
Github user rxin commented on a diff in the pull request:
https://github.com/apache/spark/pull/16887#discussion_r100552370
--- Diff: core/src/main/scala/org/apache/spark/SparkContext.scala ---
@@ -2207,20 +2207,22 @@ class SparkContext(config: SparkConf) extends
Logging
Github user rxin commented on a diff in the pull request:
https://github.com/apache/spark/pull/16887#discussion_r100552660
--- Diff: core/src/main/scala/org/apache/spark/scheduler/DAGScheduler.scala
---
@@ -696,9 +696,9 @@ class DAGScheduler(
/**
* Cancel a job that
Github user rxin commented on the issue:
https://github.com/apache/spark/pull/16664
Sorry I'm really confused, probably because I haven't kept track with this
pr. But the diff doesn't match the pr description. Are we fixing a bug here or
introducing a bunch of new
Github user rxin commented on a diff in the pull request:
https://github.com/apache/spark/pull/16664#discussion_r100564925
--- Diff:
sql/core/src/main/scala/org/apache/spark/sql/DataFrameWriter.scala ---
@@ -218,7 +247,14 @@ final class DataFrameWriter[T] private[sql](ds:
Dataset
Github user rxin commented on a diff in the pull request:
https://github.com/apache/spark/pull/16664#discussion_r100565522
--- Diff:
sql/core/src/main/scala/org/apache/spark/sql/util/QueryExecutionListener.scala
---
@@ -44,27 +44,50 @@ trait QueryExecutionListener
Github user rxin commented on a diff in the pull request:
https://github.com/apache/spark/pull/16664#discussion_r100565585
--- Diff: docs/sql-programming-guide.md ---
@@ -1300,10 +1300,28 @@ Configuration of in-memory caching can be done
using the `setConf` method on `Sp
Github user rxin commented on the issue:
https://github.com/apache/spark/pull/16887
LGTM pending Jenkins.
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so
Github user rxin commented on the issue:
https://github.com/apache/spark/pull/16885
Thanks - merging in master.
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and
Github user rxin commented on the issue:
https://github.com/apache/spark/pull/16664
That's probably because you are not familiar with the SQL component. The
existing API already has references to the QueryExecution object, which
actually includes all of the information
Github user rxin commented on the issue:
https://github.com/apache/spark/pull/16664
Well it does. It contains the entire plan.
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
Github user rxin commented on the issue:
https://github.com/apache/spark/pull/16664
I think that's a separate "bug" we should fix, i.e. DataFrameWriter should
use InsertIntoDataSourceCommand so we can consolidate the two paths.
---
If your project is set up for it, y
Github user rxin commented on the issue:
https://github.com/apache/spark/pull/16664
Basically I see no reason to add some specific parameter to a listener API
that is meant to be generic which already contains reference to QueryExecution.
What are you going to do if next time you
Github user rxin commented on the issue:
https://github.com/apache/spark/pull/16664
Actually @cloud-fan are you sure it is a problem right now?
DataSOurce.write itself creates the commands, and if the information are
propagated correctly, the QueryExecution object should have a
Github user rxin commented on the issue:
https://github.com/apache/spark/pull/16664
Yea we should fix that.
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes
Github user rxin commented on the issue:
https://github.com/apache/spark/pull/16888
Shouldn't we use netty-4.0.44.Final rather than 4.1.x?
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not
Github user rxin commented on the issue:
https://github.com/apache/spark/pull/16888
BTW for Netty we shouldn't just bump to the highest version. We should use
the maintenance branches.
---
If your project is set up for it, you can reply to this email and have your
reply appe
Github user rxin commented on the issue:
https://github.com/apache/spark/pull/16887
Merging in master!
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or
Github user rxin commented on a diff in the pull request:
https://github.com/apache/spark/pull/16386#discussion_r100687458
--- Diff:
sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/json/JacksonParser.scala
---
@@ -48,69 +47,110 @@ class JacksonParser(
// A
Github user rxin commented on the issue:
https://github.com/apache/spark/pull/16888
Are there specific benefits brought by updating to 4.1 of Netty? Netty is
so core to Spark that any bug in it would be extremely difficult to debug (yes
we have founds bugs in Netty and helped fix
Github user rxin commented on a diff in the pull request:
https://github.com/apache/spark/pull/16872#discussion_r100789955
--- Diff:
sql/core/src/test/scala/org/apache/spark/sql/DataFrameRangeSuite.scala ---
@@ -127,4 +133,28 @@ class DataFrameRangeSuite extends QueryTest with
Github user rxin commented on the issue:
https://github.com/apache/spark/pull/16914
LGTM pending jenkins.
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so
Github user rxin commented on the issue:
https://github.com/apache/spark/pull/16914
Merging in master.
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or
Github user rxin commented on the issue:
https://github.com/apache/spark/pull/14426
@dongjoon-hyun do you have time to update the pull request now the view
canonicalization work is done? Basically we can remove all the SQL generation
stuff.
---
If your project is set up for it
Github user rxin commented on the issue:
https://github.com/apache/spark/pull/14426
Actually I have some time. I will submit a pr based on this.
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not
GitHub user rxin opened a pull request:
https://github.com/apache/spark/pull/16925
[SPARK-16475][SQL] Broadcast Hint for SQL Queries
## What changes were proposed in this pull request?
This PR aims to achieve the following two goals in Spark SQL.
1. Generic Hint
Github user rxin commented on the issue:
https://github.com/apache/spark/pull/16925
Actually I'm going to completely rewrite this. I don't think the current
implementation makes sense.
---
If your project is set up for it, you can reply to this email and have your
reply
Github user rxin commented on a diff in the pull request:
https://github.com/apache/spark/pull/16925#discussion_r101088496
--- Diff:
sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/analysis/SubstituteHints.scala
---
@@ -0,0 +1,85 @@
+/*
+ * Licensed to the
Github user rxin commented on the issue:
https://github.com/apache/spark/pull/16925
cc @dongjoon-hyun, @cloud-fan, @gatorsmile and @hvanhovell This should be
ready for review. Note that the semantics is different from the earlier
versions.
---
If your project is set up for it, you
Github user rxin commented on a diff in the pull request:
https://github.com/apache/spark/pull/16925#discussion_r101129453
--- Diff:
sql/catalyst/src/test/scala/org/apache/spark/sql/catalyst/analysis/SubstituteHintsSuite.scala
---
@@ -0,0 +1,123 @@
+/*
+ * Licensed to the
Github user rxin commented on a diff in the pull request:
https://github.com/apache/spark/pull/16925#discussion_r101129594
--- Diff:
sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/analysis/SubstituteHints.scala
---
@@ -0,0 +1,103 @@
+/*
+ * Licensed to the
Github user rxin commented on a diff in the pull request:
https://github.com/apache/spark/pull/16925#discussion_r101129634
--- Diff:
sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/analysis/SubstituteHints.scala
---
@@ -0,0 +1,103 @@
+/*
+ * Licensed to the
Github user rxin commented on a diff in the pull request:
https://github.com/apache/spark/pull/16925#discussion_r101137229
--- Diff:
sql/catalyst/src/test/scala/org/apache/spark/sql/catalyst/analysis/SubstituteHintsSuite.scala
---
@@ -0,0 +1,123 @@
+/*
+ * Licensed to the
Github user rxin commented on the issue:
https://github.com/apache/spark/pull/16925
the latest commit hasn't finished running tests yet ... but probably fine
given the small change.
---
If your project is set up for it, you can reply to this email and have your
reply appe
GitHub user rxin opened a pull request:
https://github.com/apache/spark/pull/16939
[SPARK-16475][SQL] broadcast hint for SQL queries - follow up
## What changes were proposed in this pull request?
A small update to https://github.com/apache/spark/pull/16925
1. Rename
Github user rxin commented on the issue:
https://github.com/apache/spark/pull/16920
Why is this necessary? It seems like an extra step needed and doesn't
provide any real information.
I suggest you use this:
https://chrome.google.com/webstore/detail/j
Github user rxin commented on the issue:
https://github.com/apache/spark/pull/16940
LGTM (pending Jenkins).
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes
Github user rxin commented on the issue:
https://github.com/apache/spark/pull/16920
Yea the only issue is that it requires another manual update. Why not use
the chrome plugin I sent?
---
If your project is set up for it, you can reply to this email and have your
reply appear on
Github user rxin commented on a diff in the pull request:
https://github.com/apache/spark/pull/16925#discussion_r101288304
--- Diff:
sql/catalyst/src/main/antlr4/org/apache/spark/sql/catalyst/parser/SqlBase.g4 ---
@@ -374,6 +374,16 @@ querySpecification
windows
Github user rxin commented on a diff in the pull request:
https://github.com/apache/spark/pull/16925#discussion_r101289574
--- Diff:
sql/catalyst/src/main/antlr4/org/apache/spark/sql/catalyst/parser/SqlBase.g4 ---
@@ -374,6 +374,16 @@ querySpecification
windows
Github user rxin commented on a diff in the pull request:
https://github.com/apache/spark/pull/16925#discussion_r101289645
--- Diff:
sql/catalyst/src/main/antlr4/org/apache/spark/sql/catalyst/parser/SqlBase.g4 ---
@@ -374,6 +374,16 @@ querySpecification
windows
Github user rxin commented on the issue:
https://github.com/apache/spark/pull/16940
Merging in master.
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or
Github user rxin commented on a diff in the pull request:
https://github.com/apache/spark/pull/16941#discussion_r101329235
--- Diff:
sql/catalyst/src/test/scala/org/apache/spark/sql/catalyst/parser/PlanParserSuite.scala
---
@@ -524,7 +530,7 @@ class PlanParserSuite extends
Github user rxin commented on the issue:
https://github.com/apache/spark/pull/16941
Merging in master.
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or
Github user rxin commented on the issue:
https://github.com/apache/spark/pull/16943
Merging in master.
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or
Github user rxin commented on a diff in the pull request:
https://github.com/apache/spark/pull/16956#discussion_r101530187
--- Diff:
sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/analysis/ResolveHints.scala
---
@@ -54,10 +54,6 @@ object ResolveHints
GitHub user rxin opened a pull request:
https://github.com/apache/spark/pull/16958
[SPARK-13721][SQL] Make GeneratorOuter unresolved.
## What changes were proposed in this pull request?
This is a small change to make GeneratorOuter always unresolved. It is
mostly no-op change
Github user rxin commented on the issue:
https://github.com/apache/spark/pull/16958
cc @hvanhovell @bogdanrdc
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes
Github user rxin commented on the issue:
https://github.com/apache/spark/pull/16534
Change looks good to me but I didn't look super carefully.
@holdenk can you take a look at this?
---
If your project is set up for it, you can reply to this email and have your
reply a
Github user rxin commented on a diff in the pull request:
https://github.com/apache/spark/pull/16611#discussion_r101553890
--- Diff:
sql/core/src/main/scala/org/apache/spark/sql/DataFrameReader.scala ---
@@ -97,6 +99,15 @@ class DataFrameReader private[sql](sparkSession
Github user rxin commented on the issue:
https://github.com/apache/spark/pull/16611
For SQL, rather than "array", can we follow Python, e.g.
```
CREATE TEMPORARY TABLE tableA USING csv
OPTIONS (nullValue ['NA', 'null'], ...)
```
---
Github user rxin commented on the issue:
https://github.com/apache/spark/pull/16826
What's WIP about this?
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wish
Github user rxin commented on the issue:
https://github.com/apache/spark/pull/16958
So nice when I got two LGTMs and then Jenkins disagreed.
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have
Github user rxin commented on a diff in the pull request:
https://github.com/apache/spark/pull/16960#discussion_r101575199
--- Diff:
sql/core/src/test/scala/org/apache/spark/sql/execution/metric/SQLMetricsSuite.scala
---
@@ -309,4 +314,84 @@ class SQLMetricsSuite extends
Github user rxin commented on a diff in the pull request:
https://github.com/apache/spark/pull/16960#discussion_r101575264
--- Diff:
sql/core/src/test/scala/org/apache/spark/sql/execution/metric/SQLMetricsSuite.scala
---
@@ -309,4 +314,84 @@ class SQLMetricsSuite extends
Github user rxin commented on the issue:
https://github.com/apache/spark/pull/16960
cc @hvanhovell if you have a min to review this ...
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this
Github user rxin commented on the issue:
https://github.com/apache/spark/pull/16960
Merging in master.
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or
Github user rxin commented on the issue:
https://github.com/apache/spark/pull/16977
Are tests flaky right now? Otherwise it seems like this has introduced
legitimate issue with the test timing out. Three times in a row.
---
If your project is set up for it, you can reply to this
GitHub user rxin opened a pull request:
https://github.com/apache/spark/pull/17002
[SPARK-19669][SQL] Open up visibility for sharedState, sessionState, and a
few other functions
## What changes were proposed in this pull request?
To ease debugging, most of Spark SQL internals
Github user rxin commented on a diff in the pull request:
https://github.com/apache/spark/pull/17002#discussion_r102070142
--- Diff: sql/core/src/main/scala/org/apache/spark/sql/SparkSession.scala
---
@@ -95,16 +95,26 @@ class SparkSession private(
/**
* State
Github user rxin commented on the issue:
https://github.com/apache/spark/pull/17002
Yea @gatorsmile be careful in the future and check the commit hash.
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project
Github user rxin commented on a diff in the pull request:
https://github.com/apache/spark/pull/17049#discussion_r102881054
--- Diff:
sql/catalyst/src/test/scala/org/apache/spark/sql/catalyst/expressions/HashExpressionsSuite.scala
---
@@ -71,6 +75,242 @@ class HashExpressionsSuite
Github user rxin commented on the issue:
https://github.com/apache/spark/pull/17049
Looks good except that comment.
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and
Github user rxin commented on a diff in the pull request:
https://github.com/apache/spark/pull/17053#discussion_r102889140
--- Diff:
sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/catalog/ExternalCatalog.scala
---
@@ -251,7 +251,8 @@ abstract class ExternalCatalog
Github user rxin commented on the issue:
https://github.com/apache/spark/pull/17049
Merging in master.
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or
Github user rxin commented on the issue:
https://github.com/apache/spark/pull/16378
cc @cloud-fan too
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or
Github user rxin commented on a diff in the pull request:
https://github.com/apache/spark/pull/16379#discussion_r93560851
--- Diff:
sql/catalyst/src/test/scala/org/apache/spark/sql/catalyst/analysis/PullOutNondeterministicSuite.scala
---
@@ -0,0 +1,56 @@
+/*
+ * Licensed
GitHub user rxin opened a pull request:
https://github.com/apache/spark/pull/16381
[SPARK-18973][SQL] Remove SortPartitions and RedistributeData
## What changes were proposed in this pull request?
SortPartitions and RedistributeData logical operators are not actually used
and
Github user rxin commented on the issue:
https://github.com/apache/spark/pull/16381
Please also merge this into branch-2.1 to minimize backport conflicts ...
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your
Github user rxin commented on a diff in the pull request:
https://github.com/apache/spark/pull/16381#discussion_r93564401
--- Diff:
sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/plans/logical/partitioning.scala
---
@@ -1,49 +0,0 @@
-/*
- * Licensed to the
Github user rxin commented on the issue:
https://github.com/apache/spark/pull/16380
LGTM.
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the
Github user rxin commented on the issue:
https://github.com/apache/spark/pull/16381
Note that this is code from the initial Spark SQL commit!
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have
Github user rxin commented on the issue:
https://github.com/apache/spark/pull/16378
I've also cherry picked this into branch-2.1.
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have
Github user rxin commented on the issue:
https://github.com/apache/spark/pull/16349
Merging in master/branch-2.1.
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and
Github user rxin commented on the issue:
https://github.com/apache/spark/pull/14627
Thanks - merging in master/branch-2.1.
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
Github user rxin commented on the issue:
https://github.com/apache/spark/pull/14627
Actually there is a conflict. Does this fix any bug? If not we don't need
to merge it in 2.1.
---
If your project is set up for it, you can reply to this email and have your
reply appear on G
Github user rxin commented on a diff in the pull request:
https://github.com/apache/spark/pull/15923#discussion_r93669807
--- Diff:
core/src/main/scala/org/apache/spark/storage/ShuffleBlockFetcherIterator.scala
---
@@ -305,40 +316,84 @@ final class ShuffleBlockFetcherIterator
Github user rxin commented on the issue:
https://github.com/apache/spark/pull/16382
Merging in master.
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or
Github user rxin commented on the issue:
https://github.com/apache/spark/pull/16382
LGTM
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature
Github user rxin commented on the issue:
https://github.com/apache/spark/pull/16371
sounds good.
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the
Github user rxin commented on a diff in the pull request:
https://github.com/apache/spark/pull/16386#discussion_r93731259
--- Diff: python/pyspark/sql/readwriter.py ---
@@ -155,21 +155,24 @@ def load(self, path=None, format=None, schema=None,
**options):
return
Github user rxin commented on the issue:
https://github.com/apache/spark/pull/16395
cc @srinathshankar
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or
Github user rxin commented on the issue:
https://github.com/apache/spark/pull/16435
What exactly is the new policy? I don't think size in bytes is a good
choice, since it's the number of objects that can destroy GC.
---
If your project is set up for it, you can reply to
Github user rxin commented on the issue:
https://github.com/apache/spark/pull/16308
@hvanhovell anything else to do here other than bringing it up to date?
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project
Github user rxin commented on a diff in the pull request:
https://github.com/apache/spark/pull/16401#discussion_r94708121
--- Diff:
sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/plans/logical/LogicalPlan.scala
---
@@ -95,6 +96,29 @@ abstract class LogicalPlan extends
Github user rxin commented on the issue:
https://github.com/apache/spark/pull/16347
Maybe we should make DataFrameWriter.sortBy work here.
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have
Github user rxin commented on a diff in the pull request:
https://github.com/apache/spark/pull/16401#discussion_r94731500
--- Diff:
sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/plans/logical/LogicalPlan.scala
---
@@ -95,6 +96,29 @@ abstract class LogicalPlan extends
Github user rxin commented on a diff in the pull request:
https://github.com/apache/spark/pull/16401#discussion_r94732192
--- Diff:
sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/plans/logical/LogicalPlan.scala
---
@@ -95,6 +96,29 @@ abstract class LogicalPlan extends
Github user rxin commented on the issue:
https://github.com/apache/spark/pull/16475
Can we please close this?
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes
Github user rxin commented on the issue:
https://github.com/apache/spark/pull/16347
What I was suggesting was to allow sort by without bucketing.
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not
Github user rxin commented on a diff in the pull request:
https://github.com/apache/spark/pull/16401#discussion_r94859595
--- Diff:
sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/plans/logical/LogicalPlan.scala
---
@@ -95,6 +96,29 @@ abstract class LogicalPlan extends
Github user rxin commented on the issue:
https://github.com/apache/spark/pull/16337
Go for it!
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the
Github user rxin commented on a diff in the pull request:
https://github.com/apache/spark/pull/16474#discussion_r94892001
--- Diff:
sql/core/src/main/scala/org/apache/spark/sql/execution/datasources/parquet/ParquetFileFormat.scala
---
@@ -593,13 +650,10 @@ object
Github user rxin commented on a diff in the pull request:
https://github.com/apache/spark/pull/16474#discussion_r94892533
--- Diff:
sql/core/src/main/scala/org/apache/spark/sql/execution/datasources/parquet/ParquetFileFormat.scala
---
@@ -593,13 +650,10 @@ object
Github user rxin commented on a diff in the pull request:
https://github.com/apache/spark/pull/16474#discussion_r94892571
--- Diff:
sql/core/src/main/scala/org/apache/spark/sql/execution/datasources/parquet/ParquetFileFormat.scala
---
@@ -543,6 +546,58 @@ object ParquetFileFormat
Github user rxin commented on a diff in the pull request:
https://github.com/apache/spark/pull/16430#discussion_r94901694
--- Diff:
sql/core/src/test/scala/org/apache/spark/sql/estimation/EstimationSuite.scala
---
@@ -0,0 +1,81 @@
+/*
+ * Licensed to the Apache Software
Github user rxin commented on a diff in the pull request:
https://github.com/apache/spark/pull/16430#discussion_r94902084
--- Diff:
sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/plans/logical/estimation/EstimationUtils.scala
---
@@ -0,0 +1,54
Github user rxin commented on a diff in the pull request:
https://github.com/apache/spark/pull/16430#discussion_r94902048
--- Diff:
sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/plans/logical/estimation/EstimationUtils.scala
---
@@ -0,0 +1,54
1601 - 1700 of 15085 matches
Mail list logo