GitHub user bkpathak reopened a pull request:
https://github.com/apache/spark/pull/12691
[Spark-14761][SQL][WIP] Reject invalid join methods when join columns are
not specified in PySpark DataFrame join.
## What changes were proposed in this pull request?
In PySpark, the invalid join type will not throw error for the following
join:
```df1.join(df2, how='not-a-valid-join-type')```
The signature of the join is:
```def join(self, other, on=None, how=None):```
The existing code completely ignores the `how` parameter when `on` is
`None`. This patch will process the arguments passed to join and pass in to JVM
Spark SQL Analyzer, which will validate the join type passed.
## How was this patch tested?
Used manual and existing test suites.
You can merge this pull request into a Git repository by running:
$ git pull https://github.com/bkpathak/spark spark-14761
Alternatively you can review and apply these changes as the patch at:
https://github.com/apache/spark/pull/12691.patch
To close this pull request, make a commit to your master/trunk branch
with (at least) the following in the commit message:
This closes #12691
----
commit c76baff0cc4775c2191d075cc9a8176e4915fec8
Author: Bryan Cutler <[email protected]>
Date: 2016-09-11T09:19:39Z
[SPARK-17336][PYSPARK] Fix appending multiple times to PYTHONPATH from
spark-config.sh
## What changes were proposed in this pull request?
During startup of Spark standalone, the script file spark-config.sh appends
to the PYTHONPATH and can be sourced many times, causing duplicates in the
path. This change adds a env flag that is set when the PYTHONPATH is appended
so it will happen only one time.
## How was this patch tested?
Manually started standalone master/worker and verified PYTHONPATH has no
duplicate entries.
Author: Bryan Cutler <[email protected]>
Closes #15028 from BryanCutler/fix-duplicate-pythonpath-SPARK-17336.
commit 883c7631847a95684534222c1b6cfed8e62710c8
Author: Yanbo Liang <[email protected]>
Date: 2016-09-11T12:47:13Z
[SPARK-17389][FOLLOW-UP][ML] Change KMeans k-means|| default init steps
from 5 to 2.
## What changes were proposed in this pull request?
#14956 reduced default k-means|| init steps to 2 from 5 only for
spark.mllib package, we should also do same change for spark.ml and PySpark.
## How was this patch tested?
Existing tests.
Author: Yanbo Liang <[email protected]>
Closes #15050 from yanboliang/spark-17389.
commit 767d48076971f6f1e2c93ee540a9b2e5e465631b
Author: Sameer Agarwal <[email protected]>
Date: 2016-09-11T15:35:27Z
[SPARK-17415][SQL] Better error message for driver-side broadcast join OOMs
## What changes were proposed in this pull request?
This is a trivial patch that catches all `OutOfMemoryError` while building
the broadcast hash relation and rethrows it by wrapping it in a nice error
message.
## How was this patch tested?
Existing Tests
Author: Sameer Agarwal <[email protected]>
Closes #14979 from sameeragarwal/broadcast-join-error.
commit 72eec70bdbf6fb67c977463db5d8d95dd3040ae8
Author: Josh Rosen <[email protected]>
Date: 2016-09-12T04:51:22Z
[SPARK-17486] Remove unused TaskMetricsUIData.updatedBlockStatuses field
The `TaskMetricsUIData.updatedBlockStatuses` field is assigned to but never
read, increasing the memory consumption of the web UI. We should remove this
field.
Author: Josh Rosen <[email protected]>
Closes #15038 from
JoshRosen/remove-updated-block-statuses-from-TaskMetricsUIData.
commit cc87280fcd065b01667ca7a59a1a32c7ab757355
Author: cenyuhai <[email protected]>
Date: 2016-09-12T10:52:56Z
[SPARK-17171][WEB UI] DAG will list all partitions in the graph
## What changes were proposed in this pull request?
DAG will list all partitions in the graph, it is too slow and hard to see
all graph.
Always we don't want to see all partitionsï¼we just want to see the
relations of DAG graph.
So I just show 2 root nodes for Rdds.
Before this PR, the DAG graph looks like
[dag1.png](https://issues.apache.org/jira/secure/attachment/12824702/dag1.png),
[dag3.png](https://issues.apache.org/jira/secure/attachment/12825456/dag3.png),
after this PR, the DAG graph looks like
[dag2.png](https://issues.apache.org/jira/secure/attachment/12824703/dag2.png),[dag4.png](https://issues.apache.org/jira/secure/attachment/12825457/dag4.png)
Author: cenyuhai <[email protected]>
Author: å²çæµ· <[email protected]>
Closes #14737 from cenyuhai/SPARK-17171.
commit 4efcdb7feae24e41d8120b59430f8b77cc2106a6
Author: codlife <[email protected]>
Date: 2016-09-12T11:10:46Z
[SPARK-17447] Performance improvement in Partitioner.defaultPartitioner
without sortBy
## What changes were proposed in this pull request?
if there are many rdds in some situations,the sort will loss he performance
servely,actually we needn't sort the rdds , we can just scan the rdds one time
to gain the same goal.
## How was this patch tested?
manual tests
Author: codlife <[email protected]>
Closes #15039 from codlife/master.
commit b3c22912284c2a010a4af3c43dc5e6fd53c68f8c
Author: Gaetan Semet <[email protected]>
Date: 2016-09-12T11:21:33Z
[SPARK-16992][PYSPARK] use map comprehension in doc
Code is equivalent, but map comprehency is most of the time faster than a
map.
Author: Gaetan Semet <[email protected]>
Closes #14863 from Stibbons/map_comprehension.
commit 8087ecf8daad1587d0ce9040991b14320628a65e
Author: WeichenXu <[email protected]>
Date: 2016-09-12T11:23:16Z
[SPARK CORE][MINOR] fix "default partitioner cannot partition array keys"
error message in PairRDDfunctions
## What changes were proposed in this pull request?
In order to avoid confusing user,
error message in `PairRDDfunctions`
`Default partitioner cannot partition array keys.`
is updated,
the one in `partitionBy` is replaced with
`Specified partitioner cannot partition array keys.`
other is replaced with
`Specified or default partitioner cannot partition array keys.`
## How was this patch tested?
N/A
Author: WeichenXu <[email protected]>
Closes #15045 from WeichenXu123/fix_partitionBy_error_message.
commit 1742c3ab86d75ce3d352f7cddff65e62fb7c8dd4
Author: Sean Zhong <[email protected]>
Date: 2016-09-12T18:30:06Z
[SPARK-17503][CORE] Fix memory leak in Memory store when unable to cache
the whole RDD in memory
## What changes were proposed in this pull request?
MemoryStore may throw OutOfMemoryError when trying to cache a super big
RDD that cannot fit in memory.
```
scala> sc.parallelize(1 to 1000000000, 100).map(x => new
Array[Long](1000)).cache().count()
java.lang.OutOfMemoryError: Java heap space
at
$line14.$read$$iw$$iw$$iw$$iw$$iw$$iw$$iw$$iw$$anonfun$1.apply(<console>:24)
at
$line14.$read$$iw$$iw$$iw$$iw$$iw$$iw$$iw$$iw$$anonfun$1.apply(<console>:23)
at scala.collection.Iterator$$anon$11.next(Iterator.scala:409)
at scala.collection.Iterator$JoinIterator.next(Iterator.scala:232)
at
org.apache.spark.storage.memory.PartiallyUnrolledIterator.next(MemoryStore.scala:683)
at
org.apache.spark.InterruptibleIterator.next(InterruptibleIterator.scala:43)
at org.apache.spark.util.Utils$.getIteratorSize(Utils.scala:1684)
at org.apache.spark.rdd.RDD$$anonfun$count$1.apply(RDD.scala:1134)
at org.apache.spark.rdd.RDD$$anonfun$count$1.apply(RDD.scala:1134)
at
org.apache.spark.SparkContext$$anonfun$runJob$5.apply(SparkContext.scala:1915)
at
org.apache.spark.SparkContext$$anonfun$runJob$5.apply(SparkContext.scala:1915)
at org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:70)
at org.apache.spark.scheduler.Task.run(Task.scala:86)
at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:274)
at
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
at
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
at java.lang.Thread.run(Thread.java:745)
```
Spark MemoryStore uses SizeTrackingVector as a temporary unrolling buffer
to store all input values that it has read so far before transferring the
values to storage memory cache. The problem is that when the input RDD is too
big for caching in memory, the temporary unrolling memory SizeTrackingVector is
not garbage collected in time. As SizeTrackingVector can occupy all available
storage memory, it may cause the executor JVM to run out of memory quickly.
More info can be found at https://issues.apache.org/jira/browse/SPARK-17503
## How was this patch tested?
Unit test and manual test.
### Before change
Heap memory consumption
<img width="702" alt="screen shot 2016-09-12 at 4 16 15 pm"
src="https://cloud.githubusercontent.com/assets/2595532/18429524/60d73a26-7906-11e6-9768-6f286f5c58c8.png">
Heap dump
<img width="1402" alt="screen shot 2016-09-12 at 4 34 19 pm"
src="https://cloud.githubusercontent.com/assets/2595532/18429577/cbc1ef20-7906-11e6-847b-b5903f450b3b.png">
### After change
Heap memory consumption
<img width="706" alt="screen shot 2016-09-12 at 4 29 10 pm"
src="https://cloud.githubusercontent.com/assets/2595532/18429503/4abe9342-7906-11e6-844a-b2f815072624.png">
Author: Sean Zhong <[email protected]>
Closes #15056 from clockfly/memory_store_leak.
commit 3d40896f410590c0be044b3fa7e5d32115fac05e
Author: Josh Rosen <[email protected]>
Date: 2016-09-12T20:09:33Z
[SPARK-17483] Refactoring in BlockManager status reporting and block removal
This patch makes three minor refactorings to the BlockManager:
- Move the `if (info.tellMaster)` check out of `reportBlockStatus`; this
fixes an issue where a debug logging message would incorrectly claim to have
reported a block status to the master even though no message had been sent (in
case `info.tellMaster == false`). This also makes it easier to write code which
unconditionally sends block statuses to the master (which is necessary in
another patch of mine).
- Split `removeBlock()` into two methods, the existing method and an
internal `removeBlockInternal()` method which is designed to be called by
internal code that already holds a write lock on the block. This is also needed
by a followup patch.
- Instead of calling `getCurrentBlockStatus()` in `removeBlock()`, just
pass `BlockStatus.empty`; the block status should always be empty following
complete removal of a block.
These changes were originally authored as part of a bug fix patch which is
targeted at branch-2.0 and master; I've split them out here into their own
separate PR in order to make them easier to review and so that the
behavior-changing parts of my other patch can be isolated to their own PR.
Author: Josh Rosen <[email protected]>
Closes #15036 from
JoshRosen/cache-failure-race-conditions-refactorings-only.
commit 7c51b99a428a965ff7d136e1cdda20305d260453
Author: Josh Rosen <[email protected]>
Date: 2016-09-12T22:24:33Z
[SPARK-14818] Post-2.0 MiMa exclusion and build changes
This patch makes a handful of post-Spark-2.0 MiMa exclusion and build
updates. It should be merged to master and a subset of it should be picked into
branch-2.0 in order to test Spark 2.0.1-SNAPSHOT.
- Remove the ` sketch`, `mllibLocal`, and `streamingKafka010` from the list
of excluded subprojects so that MiMa checks them.
- Remove now-unnecessary special-case handling of the Kafka 0.8 artifact in
`mimaSettings`.
- Move the exclusion added in SPARK-14743 from `v20excludes` to
`v21excludes`, since that patch was only merged into master and not branch-2.0.
- Add exclusions for an API change introduced by SPARK-17096 / #14675.
- Add missing exclusions for the `o.a.spark.internal` and
`o.a.spark.sql.internal` packages.
Author: Josh Rosen <[email protected]>
Closes #15061 from JoshRosen/post-2.0-mima-changes.
commit f9c580f11098d95f098936a0b90fa21d71021205
Author: Josh Rosen <[email protected]>
Date: 2016-09-12T22:43:57Z
[SPARK-17485] Prevent failed remote reads of cached blocks from failing
entire job
## What changes were proposed in this pull request?
In Spark's `RDD.getOrCompute` we first try to read a local copy of a cached
RDD block, then a remote copy, and only fall back to recomputing the block if
no cached copy (local or remote) can be read. This logic works correctly in the
case where no remote copies of the block exist, but if there _are_ remote
copies and reads of those copies fail (due to network issues or internal Spark
bugs) then the BlockManager will throw a `BlockFetchException` that will fail
the task (and which could possibly fail the whole job if the read failures keep
occurring).
In the cases of TorrentBroadcast and task result fetching we really do want
to fail the entire job in case no remote blocks can be fetched, but this logic
is inappropriate for reads of cached RDD blocks because those can/should be
recomputed in case cached blocks are unavailable.
Therefore, I think that the `BlockManager.getRemoteBytes()` method should
never throw on remote fetch errors and, instead, should handle failures by
returning `None`.
## How was this patch tested?
Block manager changes should be covered by modified tests in
`BlockManagerSuite`: the old tests expected exceptions to be thrown on failed
remote reads, while the modified tests now expect `None` to be returned from
the `getRemote*` method.
I also manually inspected all usages of `BlockManager.getRemoteValues()`,
`getRemoteBytes()`, and `get()` to verify that they correctly pattern-match on
the result and handle `None`. Note that these `None` branches are already
exercised because the old `getRemoteBytes` returned `None` when no remote
locations for the block could be found (which could occur if an executor died
and its block manager de-registered with the master).
Author: Josh Rosen <[email protected]>
Closes #15037 from JoshRosen/SPARK-17485.
commit a91ab705e8c124aa116c3e5b1f3ba88ce832dcde
Author: Davies Liu <[email protected]>
Date: 2016-09-12T23:35:42Z
[SPARK-17474] [SQL] fix python udf in TakeOrderedAndProjectExec
## What changes were proposed in this pull request?
When there is any Python UDF in the Project between Sort and Limit, it will
be collected into TakeOrderedAndProjectExec, ExtractPythonUDFs failed to pull
the Python UDFs out because QueryPlan.expressions does not include the
expression inside Option[Seq[Expression]].
Ideally, we should fix the `QueryPlan.expressions`, but tried with no luck
(it always run into infinite loop). In PR, I changed the
TakeOrderedAndProjectExec to no use Option[Seq[Expression]] to workaround it.
cc JoshRosen
## How was this patch tested?
Added regression test.
Author: Davies Liu <[email protected]>
Closes #15030 from davies/all_expr.
commit 46f5c201e70053635bdeab4984ba1b649478bd12
Author: hyukjinkwon <[email protected]>
Date: 2016-09-13T09:42:51Z
[BUILD] Closing some stale PRs and ones suggested to be closed by
committer(s)
## What changes were proposed in this pull request?
This PR proposes to close some stale PRs and ones suggested to be closed by
committer(s)
Closes #10052
Closes #11079
Closes #12661
Closes #12772
Closes #12958
Closes #12990
Closes #13409
Closes #13779
Closes #13811
Closes #14577
Closes #14714
Closes #14875
Closes #15020
## How was this patch tested?
N/A
Author: hyukjinkwon <[email protected]>
Closes #15057 from HyukjinKwon/closing-stale-pr.
commit 3f6a2bb3f7beac4ce928eb660ee36258b5b9e8c8
Author: Josh Rosen <[email protected]>
Date: 2016-09-13T10:54:03Z
[SPARK-17515] CollectLimit.execute() should perform per-partition limits
## What changes were proposed in this pull request?
CollectLimit.execute() incorrectly omits per-partition limits, leading to
performance regressions in case this case is hit (which should not happen in
normal operation, but can occur in some cases (see #15068 for one example).
## How was this patch tested?
Regression test in SQLQuerySuite that asserts the number of records scanned
from the input RDD.
Author: Josh Rosen <[email protected]>
Closes #15070 from JoshRosen/SPARK-17515.
commit 4ba63b193c1ac292493e06343d9d618c12c5ef3f
Author: jiangxingbo <[email protected]>
Date: 2016-09-13T15:04:51Z
[SPARK-17142][SQL] Complex query triggers binding error in HashAggregateExec
## What changes were proposed in this pull request?
In `ReorderAssociativeOperator` rule, we extract foldable expressions with
Add/Multiply arithmetics, and replace with eval literal. For example, `(a + 1)
+ (b + 2)` is optimized to `(a + b + 3)` by this rule.
For aggregate operator, output expressions should be derived from
groupingExpressions, current implemenation of `ReorderAssociativeOperator` rule
may break this promise. A instance could be:
```
SELECT
((t1.a + 1) + (t2.a + 2)) AS out_col
FROM
testdata2 AS t1
INNER JOIN
testdata2 AS t2
ON
(t1.a = t2.a)
GROUP BY (t1.a + 1), (t2.a + 2)
```
`((t1.a + 1) + (t2.a + 2))` is optimized to `(t1.a + t2.a + 3)`, which
could not be derived from `ExpressionSet((t1.a +1), (t2.a + 2))`.
Maybe we should improve the rule of `ReorderAssociativeOperator` by adding
a GroupingExpressionSet to keep Aggregate.groupingExpressions, and respect
these expressions during the optimize stage.
## How was this patch tested?
Add new test case in `ReorderAssociativeOperatorSuite`.
Author: jiangxingbo <[email protected]>
Closes #14917 from jiangxb1987/rao.
commit 72edc7e958271cedb01932880550cfc2c0631204
Author: Burak Yavuz <[email protected]>
Date: 2016-09-13T22:11:55Z
[SPARK-17531] Don't initialize Hive Listeners for the Execution Client
## What changes were proposed in this pull request?
If a user provides listeners inside the Hive Conf, the configuration for
these listeners are passed to the Hive Execution Client as well. This may cause
issues for two reasons:
1. The Execution Client will actually generate garbage
2. The listener class needs to be both in the Spark Classpath and Hive
Classpath
This PR empties the listener configurations in
`HiveUtils.newTemporaryConfiguration` so that the execution client will not
contain the listener confs, but the metadata client will.
## How was this patch tested?
Unit tests
Author: Burak Yavuz <[email protected]>
Closes #15086 from brkyvz/null-listeners.
commit 37b93f54e89332b6b77bb02c1c2299614338fd7c
Author: gatorsmile <[email protected]>
Date: 2016-09-13T22:37:42Z
[SPARK-17530][SQL] Add Statistics into DESCRIBE FORMATTED
### What changes were proposed in this pull request?
Statistics is missing in the output of `DESCRIBE FORMATTED`. This PR is to
add it. After the PR, the output will be like:
```
+----------------------------+----------------------------------------------------------------------------------------------------------------------+-------+
|col_name |data_type
|comment|
+----------------------------+----------------------------------------------------------------------------------------------------------------------+-------+
|key |string
|null
|
|value |string
|null
|
| |
|
|
|# Detailed Table Information|
|
|
|Database: |default
|
|
|Owner: |xiaoli
|
|
|Create Time: |Tue Sep 13 14:36:57 PDT 2016
|
|
|Last Access Time: |Wed Dec 31 16:00:00 PST 1969
|
|
|Location:
|file:/private/var/folders/4b/sgmfldk15js406vk7lw5llzw0000gn/T/warehouse-9982e1db-df17-4376-a140-dbbee0203d83/texttable|
|
|Table Type: |MANAGED
|
|
|Statistics: |sizeInBytes=5812, rowCount=500,
isBroadcastable=false
| |
|Table Parameters: |
|
|
| rawDataSize |-1
|
|
| numFiles |1
|
|
| transient_lastDdlTime |1473802620
|
|
| totalSize |5812
|
|
| COLUMN_STATS_ACCURATE |false
|
|
| numRows |-1
|
|
| |
|
|
|# Storage Information |
|
|
|SerDe Library:
|org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
| |
|InputFormat: |org.apache.hadoop.mapred.TextInputFormat
|
|
|OutputFormat:
|org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
| |
|Compressed: |No
|
|
|Storage Desc Parameters: |
|
|
| serialization.format |1
|
|
+----------------------------+----------------------------------------------------------------------------------------------------------------------+-------+
```
Also improve the output of statistics in `DESCRIBE EXTENDED` by removing
duplicate `Statistics`. Below is the example after the PR:
```
+----------------------------+-------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+-------+
|col_name |data_type
|comment|
+----------------------------+-------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+-------+
|key |string
|null |
|value |string
|null |
| |
| |
|# Detailed Table Information|CatalogTable(
Table: `default`.`texttable`
Owner: xiaoli
Created: Tue Sep 13 14:38:43 PDT 2016
Last Access: Wed Dec 31 16:00:00 PST 1969
Type: MANAGED
Schema: [StructField(key,StringType,true),
StructField(value,StringType,true)]
Provider: hive
Properties: [rawDataSize=-1, numFiles=1,
transient_lastDdlTime=1473802726, totalSize=5812, COLUMN_STATS_ACCURATE=false,
numRows=-1]
Statistics: sizeInBytes=5812, rowCount=500, isBroadcastable=false
Storage(Location:
file:/private/var/folders/4b/sgmfldk15js406vk7lw5llzw0000gn/T/warehouse-8ea5c5a0-5680-4778-91cb-c6334cf8a708/texttable,
InputFormat: org.apache.hadoop.mapred.TextInputFormat, OutputFormat:
org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat, Serde:
org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe, Properties:
[serialization.format=1]))| |
+----------------------------+-------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+-------+
```
### How was this patch tested?
Manually tested.
Author: gatorsmile <[email protected]>
Closes #15083 from gatorsmile/descFormattedStats.
commit a454a4d86bbed1b6988da0a0e23b3e87a1a16340
Author: junyangq <[email protected]>
Date: 2016-09-14T04:01:03Z
[SPARK-17317][SPARKR] Add SparkR vignette
## What changes were proposed in this pull request?
This PR tries to add a SparkR vignette, which works as a friendly guidance
going through the functionality provided by SparkR.
## How was this patch tested?
Manual test.
Author: junyangq <[email protected]>
Author: Shivaram Venkataraman <[email protected]>
Author: Junyang Qian <[email protected]>
Closes #14980 from junyangq/SPARKR-vignette.
commit def7c265f539f3e119f068b6e9050300d05b14a4
Author: Jagadeesan <[email protected]>
Date: 2016-09-14T08:03:16Z
[SPARK-17449][DOCUMENTATION] Relation between heartbeatInterval andâ¦
## What changes were proposed in this pull request?
The relation between spark.network.timeout and
spark.executor.heartbeatInterval should be mentioned in the document.
⦠network timeout]
Author: Jagadeesan <[email protected]>
Closes #15042 from jagadeesanas2/SPARK-17449.
commit b5bfcddbfbc2e79d3d0fbd43942716946e6c4ba3
Author: Sami Jaktholm <[email protected]>
Date: 2016-09-14T08:38:30Z
[SPARK-17525][PYTHON] Remove SparkContext.clearFiles() from the PySpark API
as it was removed from the Scala API prior to Spark 2.0.0
## What changes were proposed in this pull request?
This pull request removes the SparkContext.clearFiles() method from the
PySpark API as the method was removed from the Scala API in
8ce645d4eeda203cf5e100c4bdba2d71edd44e6a. Using that method in PySpark leads to
an exception as PySpark tries to call the non-existent method on the JVM side.
## How was this patch tested?
Existing tests (though none of them tested this particular method).
Author: Sami Jaktholm <[email protected]>
Closes #15081 from sjakthol/pyspark-sc-clearfiles.
commit 18b4f035f40359b3164456d0dab52dbc762ea3b4
Author: [email protected] <[email protected]>
Date: 2016-09-14T08:49:15Z
[CORE][DOC] remove redundant comment
## What changes were proposed in this pull request?
In the comment, there is redundant `the estimated`.
This PR simply remove the redundant comment and adjusts format.
Author: [email protected] <[email protected]>
Closes #15091 from wangmiao1981/comment.
commit 4cea9da2ae88b40a5503111f8f37051e2372163e
Author: Ergin Seyfe <[email protected]>
Date: 2016-09-14T08:51:14Z
[SPARK-17480][SQL] Improve performance by removing or caching List.length
which is O(n)
## What changes were proposed in this pull request?
Scala's List.length method is O(N) and it makes the
gatherCompressibilityStats function O(N^2). Eliminate the List.length calls by
writing it in Scala way.
https://github.com/scala/scala/blob/2.10.x/src/library/scala/collection/LinearSeqOptimized.scala#L36
As suggested. Extended the fix to HiveInspectors and AggregationIterator
classes as well.
## How was this patch tested?
Profiled a Spark job and found that CompressibleColumnBuilder is using 39%
of the CPU. Out of this 39%
CompressibleColumnBuilder->gatherCompressibilityStats is using 23% of it. 6.24%
of the CPU is spend on List.length which is called inside
gatherCompressibilityStats.
After this change we started to save 6.24% of the CPU.
Author: Ergin Seyfe <[email protected]>
Closes #15032 from seyfe/gatherCompressibilityStats.
commit dc0a4c916151c795dc41b5714e9d23b4937f4636
Author: Sean Owen <[email protected]>
Date: 2016-09-14T09:10:16Z
[SPARK-17445][DOCS] Reference an ASF page as the main place to find
third-party packages
## What changes were proposed in this pull request?
Point references to spark-packages.org to
https://cwiki.apache.org/confluence/display/SPARK/Third+Party+Projects
This will be accompanied by a parallel change to the spark-website repo,
and additional changes to this wiki.
## How was this patch tested?
Jenkins tests.
Author: Sean Owen <[email protected]>
Closes #15075 from srowen/SPARK-17445.
commit 52738d4e099a19466ef909b77c24cab109548706
Author: gatorsmile <[email protected]>
Date: 2016-09-14T15:10:20Z
[SPARK-17409][SQL] Do Not Optimize Query in CTAS More Than Once
### What changes were proposed in this pull request?
As explained in https://github.com/apache/spark/pull/14797:
>Some analyzer rules have assumptions on logical plans, optimizer may break
these assumption, we should not pass an optimized query plan into
QueryExecution (will be analyzed again), otherwise we may some weird bugs.
For example, we have a rule for decimal calculation to promote the
precision before binary operations, use PromotePrecision as placeholder to
indicate that this rule should not apply twice. But a Optimizer rule will
remove this placeholder, that break the assumption, then the rule applied
twice, cause wrong result.
We should not optimize the query in CTAS more than once. For example,
```Scala
spark.range(99, 101).createOrReplaceTempView("tab1")
val sqlStmt = "SELECT id, cast(id as long) * cast('1.0' as decimal(38, 18))
as num FROM tab1"
sql(s"CREATE TABLE tab2 USING PARQUET AS $sqlStmt")
checkAnswer(spark.table("tab2"), sql(sqlStmt))
```
Before this PR, the results do not match
```
== Results ==
!== Correct Answer - 2 == == Spark Answer - 2 ==
![100,100.000000000000000000] [100,null]
[99,99.000000000000000000] [99,99.000000000000000000]
```
After this PR, the results match.
```
+---+----------------------+
|id |num |
+---+----------------------+
|99 |99.000000000000000000 |
|100|100.000000000000000000|
+---+----------------------+
```
In this PR, we do not treat the `query` in CTAS as a child. Thus, the
`query` will not be optimized when optimizing CTAS statement. However, we still
need to analyze it for normalizing and verifying the CTAS in the Analyzer.
Thus, we do it in the analyzer rule `PreprocessDDL`, because so far only this
rule needs the analyzed plan of the `query`.
### How was this patch tested?
Added a test
Author: gatorsmile <[email protected]>
Closes #15048 from gatorsmile/ctasOptimized.
commit 6d06ff6f7e2dd72ba8fe96cd875e83eda6ebb2a9
Author: Josh Rosen <[email protected]>
Date: 2016-09-14T17:10:01Z
[SPARK-17514] df.take(1) and df.limit(1).collect() should perform the same
in Python
## What changes were proposed in this pull request?
In PySpark, `df.take(1)` runs a single-stage job which computes only one
partition of the DataFrame, while `df.limit(1).collect()` computes all
partitions and runs a two-stage job. This difference in performance is
confusing.
The reason why `limit(1).collect()` is so much slower is that `collect()`
internally maps to `df.rdd.<some-pyspark-conversions>.toLocalIterator`, which
causes Spark SQL to build a query where a global limit appears in the middle of
the plan; this, in turn, ends up being executed inefficiently because limits in
the middle of plans are now implemented by repartitioning to a single task
rather than by running a `take()` job on the driver (this was done in #7334, a
patch which was a prerequisite to allowing partition-local limits to be pushed
beneath unions, etc.).
In order to fix this performance problem I think that we should generalize
the fix from SPARK-10731 / #8876 so that `DataFrame.collect()` also delegates
to the Scala implementation and shares the same performance properties. This
patch modifies `DataFrame.collect()` to first collect all results to the driver
and then pass them to Python, allowing this query to be planned using Spark's
`CollectLimit` optimizations.
## How was this patch tested?
Added a regression test in `sql/tests.py` which asserts that the expected
number of jobs, stages, and tasks are run for both queries.
Author: Josh Rosen <[email protected]>
Closes #15068 from JoshRosen/pyspark-collect-limit.
commit a79838bdeeb12cec4d50da3948bd8a33777e53a6
Author: hyukjinkwon <[email protected]>
Date: 2016-09-14T17:33:56Z
[MINOR][SQL] Add missing functions for some options in SQLConf and use them
where applicable
## What changes were proposed in this pull request?
I first thought they are missing because they are kind of hidden options
but it seems they are just missing.
For example, `spark.sql.parquet.mergeSchema` is documented in
[sql-programming-guide.md](https://github.com/apache/spark/blob/master/docs/sql-programming-guide.md)
but this function is missing whereas many options such as
`spark.sql.join.preferSortMergeJoin` are not documented but have its own
function individually.
So, this PR suggests making them consistent by adding the missing functions
for some options in `SQLConf` and use them where applicable, in order to make
them more readable.
## How was this patch tested?
Existing tests should cover this.
Author: hyukjinkwon <[email protected]>
Closes #14678 from HyukjinKwon/sqlconf-cleanup.
commit 040e46979d5f90edc7f9be3cbedd87e8986e8053
Author: Xin Wu <[email protected]>
Date: 2016-09-14T19:14:29Z
[SPARK-10747][SQL] Support NULLS FIRST|LAST clause in ORDER BY
## What changes were proposed in this pull request?
Currently, ORDER BY clause returns nulls value according to sorting order
(ASC|DESC), considering null value is always smaller than non-null values.
However, SQL2003 standard support NULLS FIRST or NULLS LAST to allow users
to specify whether null values should be returned first or last, regardless of
sorting order (ASC|DESC).
This PR is to support this new feature.
## How was this patch tested?
New test cases are added to test NULLS FIRST|LAST for regular select
queries and windowing queries.
(If this patch involves UI changes, please attach a screenshot; otherwise,
remove this)
Author: Xin Wu <[email protected]>
Closes #14842 from xwu0226/SPARK-10747.
commit ff6e4cbdc80e2ad84c5d70ee07f323fad9374e3e
Author: Kishor Patil <[email protected]>
Date: 2016-09-14T19:19:35Z
[SPARK-17511] Yarn Dynamic Allocation: Avoid marking released container as
Failed
## What changes were proposed in this pull request?
Due to race conditions, the ` assert(numExecutorsRunning <=
targetNumExecutors)` can fail causing `AssertionError`. So removed the
assertion, instead moved the conditional check before launching new container:
```
java.lang.AssertionError: assertion failed
at scala.Predef$.assert(Predef.scala:156)
at
org.apache.spark.deploy.yarn.YarnAllocator$$anonfun$runAllocatedContainers$1.org$apache$spark$deploy$yarn$YarnAllocator$$anonfun$$updateInternalState$1(YarnAllocator.scala:489)
at
org.apache.spark.deploy.yarn.YarnAllocator$$anonfun$runAllocatedContainers$1$$anon$1.run(YarnAllocator.scala:519)
at
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
at
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
at java.lang.Thread.run(Thread.java:745)
```
## How was this patch tested?
This was manually tested using a large ForkAndJoin job with Dynamic
Allocation enabled to validate the failing job succeeds, without any such
exception.
Author: Kishor Patil <[email protected]>
Closes #15069 from kishorvpatil/SPARK-17511.
commit e33bfaed3b160fbc617c878067af17477a0044f5
Author: Shixiong Zhu <[email protected]>
Date: 2016-09-14T20:33:51Z
[SPARK-17463][CORE] Make CollectionAccumulator and SetAccumulator's value
can be read thread-safely
## What changes were proposed in this pull request?
Make CollectionAccumulator and SetAccumulator's value can be read
thread-safely to fix the ConcurrentModificationException reported in
[JIRA](https://issues.apache.org/jira/browse/SPARK-17463).
## How was this patch tested?
Existing tests.
Author: Shixiong Zhu <[email protected]>
Closes #15063 from zsxwing/SPARK-17463.
----
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enabled but not working, please
contact infrastructure at [email protected] or file a JIRA ticket
with INFRA.
---
---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]