GitHub user naujlove opened a pull request:
https://github.com/apache/spark/pull/11413
Branch 1.6
## What changes were proposed in this pull request?
(Please fill in changes proposed in this fix)
## How was this patch tested?
(Please explain how this patch was tested. E.g. unit tests, integration
tests, manual tests)
(If this patch involves UI changes, please attach a screenshot; otherwise,
remove this)
You can merge this pull request into a Git repository by running:
$ git pull https://github.com/apache/spark branch-1.6
Alternatively you can review and apply these changes as the patch at:
https://github.com/apache/spark/pull/11413.patch
To close this pull request, make a commit to your master/trunk branch
with (at least) the following in the commit message:
This closes #11413
----
commit b1d5a7859546eabdc7cf070b3e78d91389a8fbd6
Author: Timothy Hunter <[email protected]>
Date: 2015-12-09T02:40:21Z
[SPARK-8517][ML][DOC] Reorganizes the spark.ml user guide
This PR moves pieces of the spark.ml user guide to reflect suggestions in
SPARK-8517. It does not introduce new content, as requested.
<img width="192" alt="screen shot 2015-12-08 at 11 36 00 am"
src="https://cloud.githubusercontent.com/assets/7594753/11666166/e82b84f2-9d9f-11e5-8904-e215424d8444.png">
Author: Timothy Hunter <[email protected]>
Closes #10207 from thunterdb/spark-8517.
(cherry picked from commit 765c67f5f2e0b1367e37883f662d313661e3a0d9)
Signed-off-by: Joseph K. Bradley <[email protected]>
commit 9e82273afc68947dc2a08315e0d42cfcedacaa2a
Author: Dominik Dahlem <[email protected]>
Date: 2015-12-09T02:54:10Z
[SPARK-11343][ML] Documentation of float and double prediction/label
columns in RegressionEvaluator
felixcheung , mengxr
Just added a message to require()
Author: Dominik Dahlem <[email protected]>
Closes #9598 from
dahlem/ddahlem_regression_evaluator_double_predictions_message_04112015.
(cherry picked from commit a0046e379bee0852c39ece4ea719cde70d350b0e)
Signed-off-by: Joseph K. Bradley <[email protected]>
commit 0be792aad5d01432e989a03969541f41a45281e2
Author: Fei Wang <[email protected]>
Date: 2015-12-09T05:32:31Z
[SPARK-12222] [CORE] Deserialize RoaringBitmap using Kryo serializer throw
Buffer underflow exception
Jira: https://issues.apache.org/jira/browse/SPARK-12222
Deserialize RoaringBitmap using Kryo serializer throw Buffer underflow
exception:
```
com.esotericsoftware.kryo.KryoException: Buffer underflow.
at com.esotericsoftware.kryo.io.Input.require(Input.java:156)
at com.esotericsoftware.kryo.io.Input.skip(Input.java:131)
at com.esotericsoftware.kryo.io.Input.skip(Input.java:264)
```
This is caused by a bug of kryo's `Input.skip(long
count)`(https://github.com/EsotericSoftware/kryo/issues/119) and we call this
method in `KryoInputDataInputBridge`.
Instead of upgrade kryo's version, this pr bypass the kryo's
`Input.skip(long count)` by directly call another `skip` method in kryo's
Input.java(https://github.com/EsotericSoftware/kryo/blob/kryo-2.21/src/com/esotericsoftware/kryo/io/Input.java#L124),
i.e. write the bug-fixed version of `Input.skip(long count)` in
KryoInputDataInputBridge's `skipBytes` method.
more detail link to
https://github.com/apache/spark/pull/9748#issuecomment-162860246
Author: Fei Wang <[email protected]>
Closes #10213 from scwf/patch-1.
(cherry picked from commit 3934562d34bbe08d91c54b4bbee27870e93d7571)
Signed-off-by: Davies Liu <[email protected]>
commit b5a76b4a40e043c5384be7c620e7ca257b7ef2cd
Author: uncleGen <[email protected]>
Date: 2015-12-09T15:09:40Z
[SPARK-12031][CORE][BUG] Integer overflow when do sampling
Author: uncleGen <[email protected]>
Closes #10023 from uncleGen/1.6-bugfix.
(cherry picked from commit a113216865fd45ea39ae8f104e784af2cf667dcf)
Signed-off-by: Sean Owen <[email protected]>
commit acd462420ab5565ba5bf098f399fb355da3d6139
Author: Holden Karau <[email protected]>
Date: 2015-12-09T16:45:13Z
[SPARK-10299][ML] word2vec should allow users to specify the window size
Currently word2vec has the window hard coded at 5, some users may want
different sizes (for example if using on n-gram input or similar). User request
comes from
http://stackoverflow.com/questions/32231975/spark-word2vec-window-size .
Author: Holden Karau <[email protected]>
Author: Holden Karau <[email protected]>
Closes #8513 from
holdenk/SPARK-10299-word2vec-should-allow-users-to-specify-the-window-size.
(cherry picked from commit 22b9a8740d51289434553d19b6b1ac34aecdc09a)
Signed-off-by: Sean Owen <[email protected]>
commit 05e441e121a86e0c105ad25010e4678f2f9e73e3
Author: Josh Rosen <[email protected]>
Date: 2015-12-09T19:39:59Z
[SPARK-12165][SPARK-12189] Fix bugs in eviction of storage memory by
execution
This patch fixes a bug in the eviction of storage memory by execution.
## The bug:
In general, execution should be able to evict storage memory when the total
storage memory usage is greater than `maxMemory *
spark.memory.storageFraction`. Due to a bug, however, Spark might wind up
evicting no storage memory in certain cases where the storage memory usage was
between `maxMemory * spark.memory.storageFraction` and `maxMemory`. For
example, here is a regression test which illustrates the bug:
```scala
val maxMemory = 1000L
val taskAttemptId = 0L
val (mm, ms) = makeThings(maxMemory)
// Since we used the default storage fraction (0.5), we should be able
to allocate 500 bytes
// of storage memory which are immune to eviction by execution memory
pressure.
// Acquire enough storage memory to exceed the storage region size
assert(mm.acquireStorageMemory(dummyBlock, 750L, evictedBlocks))
assertEvictBlocksToFreeSpaceNotCalled(ms)
assert(mm.executionMemoryUsed === 0L)
assert(mm.storageMemoryUsed === 750L)
// At this point, storage is using 250 more bytes of memory than it is
guaranteed, so execution
// should be able to reclaim up to 250 bytes of storage memory.
// Therefore, execution should now be able to require up to 500 bytes
of memory:
assert(mm.acquireExecutionMemory(500L, taskAttemptId,
MemoryMode.ON_HEAP) === 500L) // <--- fails by only returning 250L
assert(mm.storageMemoryUsed === 500L)
assert(mm.executionMemoryUsed === 500L)
assertEvictBlocksToFreeSpaceCalled(ms, 250L)
```
The problem relates to the control flow / interaction between
`StorageMemoryPool.shrinkPoolToReclaimSpace()` and
`MemoryStore.ensureFreeSpace()`. While trying to allocate the 500 bytes of
execution memory, the `UnifiedMemoryManager` discovers that it will need to
reclaim 250 bytes of memory from storage, so it calls
`StorageMemoryPool.shrinkPoolToReclaimSpace(250L)`. This method, in turn, calls
`MemoryStore.ensureFreeSpace(250L)`. However, `ensureFreeSpace()` first checks
whether the requested space is less than `maxStorageMemory -
storageMemoryUsed`, which will be true if there is any free execution memory
because it turns out that `MemoryStore.maxStorageMemory = (maxMemory -
onHeapExecutionMemoryPool.memoryUsed)` when the `UnifiedMemoryManager` is used.
The control flow here is somewhat confusing (it grew to be messy /
confusing over time / as a result of the merging / refactoring of several
components). In the pre-Spark 1.6 code, `ensureFreeSpace` was called directly
by the `MemoryStore` itself, whereas in 1.6 it's involved in a confusing
control flow where `MemoryStore` calls `MemoryManager.acquireStorageMemory`,
which then calls back into `MemoryStore.ensureFreeSpace`, which, in turn, calls
`MemoryManager.freeStorageMemory`.
## The solution:
The solution implemented in this patch is to remove the confusing circular
control flow between `MemoryManager` and `MemoryStore`, making the storage
memory acquisition process much more linear / straightforward. The key changes:
- Remove a layer of inheritance which made the memory manager code harder
to understand (53841174760a24a0df3eb1562af1f33dbe340eb9).
- Move some bounds checks earlier in the call chain
(13ba7ada77f87ef1ec362aec35c89a924e6987cb).
- Refactor `ensureFreeSpace()` so that the part which evicts blocks can be
called independently from the part which checks whether there is enough free
space to avoid eviction (7c68ca09cb1b12f157400866983f753ac863380e).
- Realize that this lets us remove a layer of overloads from
`ensureFreeSpace` (eec4f6c87423d5e482b710e098486b3bbc4daf06).
- Realize that `ensureFreeSpace()` can simply be replaced with an
`evictBlocksToFreeSpace()` method which is called [after we've already figured
out](https://github.com/apache/spark/blob/2dc842aea82c8895125d46a00aa43dfb0d121de9/core/src/main/scala/org/apache/spark/memory/StorageMemoryPool.scala#L88)
how much memory needs to be reclaimed via eviction;
(2dc842aea82c8895125d46a00aa43dfb0d121de9).
Along the way, I fixed some problems with the mocks in
`MemoryManagerSuite`: the old mocks would
[unconditionally](https://github.com/apache/spark/blob/80a824d36eec9d9a9f092ee1741453851218ec73/core/src/test/scala/org/apache/spark/memory/MemoryManagerSuite.scala#L84)
report that a block had been evicted even if there was enough space in the
storage pool such that eviction would be avoided.
I also fixed a problem where `StorageMemoryPool._memoryUsed` might become
negative due to freed memory being double-counted when excution evicts storage.
The problem was that `StorageMemoryPoolshrinkPoolToFreeSpace` would [decrement
`_memoryUsed`](https://github.com/apache/spark/commit/7c68ca09cb1b12f157400866983f753ac863380e#diff-935c68a9803be144ed7bafdd2f756a0fL133)
even though `StorageMemoryPool.freeMemory` had already decremented it as each
evicted block was freed. See SPARK-12189 for details.
Author: Josh Rosen <[email protected]>
Author: Andrew Or <[email protected]>
Closes #10170 from JoshRosen/SPARK-12165.
(cherry picked from commit aec5ea000ebb8921f42f006b694ef26f5df67d83)
Signed-off-by: Andrew Or <[email protected]>
commit ee0a6e72234e4f672a2939b794c904026f696398
Author: Sean Owen <[email protected]>
Date: 2015-12-09T19:47:38Z
[SPARK-11824][WEBUI] WebUI does not render descriptions with 'bad' HTML,
throws console error
Don't warn when description isn't valid HTML since it may properly be like
"SELECT ... where foo <= 1"
The tests for this code indicate that it's normal to handle strings like
this that don't contain HTML as a string rather than markup. Hence logging
every such instance as a warning is too noisy since it's not a problem. this is
an issue for stages whose name contain SQL like the above
CC tdas as author of this bit of code
Author: Sean Owen <[email protected]>
Closes #10159 from srowen/SPARK-11824.
(cherry picked from commit 1eb7c22ce72a1b82ed194a51bbcf0da9c771605a)
Signed-off-by: Sean Owen <[email protected]>
commit bfb4201395c6a1905c6eb46de4ea3eefe8d17309
Author: Xusen Yin <[email protected]>
Date: 2015-12-09T20:00:48Z
[SPARK-11551][DOC] Replace example code in ml-features.md using
include_example
PR on behalf of somideshmukh, thanks!
Author: Xusen Yin <[email protected]>
Author: somideshmukh <[email protected]>
Closes #10219 from yinxusen/SPARK-11551.
(cherry picked from commit 051c6a066f7b5fcc7472412144c15b50a5319bd5)
Signed-off-by: Xiangrui Meng <[email protected]>
commit 9bc6a27fdc5db3815958c721737a195af93f3757
Author: Andrew Ray <[email protected]>
Date: 2015-12-10T01:16:01Z
[SPARK-12211][DOC][GRAPHX] Fix version number in graphx doc for migration
from 1.1
Migration from 1.1 section added to the GraphX doc in 1.2.0 (see
https://spark.apache.org/docs/1.2.0/graphx-programming-guide.html#migrating-from-spark-11)
uses \{{site.SPARK_VERSION}} as the version where changes were introduced, it
should be just 1.2.
Author: Andrew Ray <[email protected]>
Closes #10206 from aray/graphx-doc-1.1-migration.
(cherry picked from commit 7a8e587dc04c2fabc875d1754eae7f85b4fba6ba)
Signed-off-by: Joseph K. Bradley <[email protected]>
commit d86a88da677041d3c4ab484ed6f4f152674091f0
Author: Andrew Or <[email protected]>
Date: 2015-12-10T01:24:04Z
[SPARK-12165][ADDENDUM] Fix outdated comments on unroll test
JoshRosen
Author: Andrew Or <[email protected]>
Closes #10229 from andrewor14/unroll-test-comments.
(cherry picked from commit 8770bd1213f9b1051dabde9c5424ae7b32143a44)
Signed-off-by: Josh Rosen <[email protected]>
commit 9fe8dc916e8a30914199b1fbb8c3765ba742559a
Author: Yin Huai <[email protected]>
Date: 2015-12-10T02:09:36Z
[SPARK-11678][SQL][DOCS] Document basePath in the programming guide.
This PR adds document for `basePath`, which is a new parameter used by
`HadoopFsRelation`.
The compiled doc is shown below.

JIRA: https://issues.apache.org/jira/browse/SPARK-11678
Author: Yin Huai <[email protected]>
Closes #10211 from yhuai/basePathDoc.
(cherry picked from commit ac8cdf1cdc148bd21290ecf4d4f9874f8c87cc14)
Signed-off-by: Yin Huai <[email protected]>
commit 699f497cf7ceefbaed689b6f3515f8a2ebc636ca
Author: Mark Grover <[email protected]>
Date: 2015-12-10T02:37:35Z
[SPARK-11796] Fix httpclient and httpcore depedency issues related to
docker-client
This commit fixes dependency issues which prevented the Docker-based JDBC
integration tests from running in the Maven build.
Author: Mark Grover <[email protected]>
Closes #9876 from markgrover/master_docker.
(cherry picked from commit 2166c2a75083c2262e071a652dd52b1a33348b6e)
Signed-off-by: Josh Rosen <[email protected]>
commit f6d8661738b5a4b139c4800d5c4e9f0094068451
Author: Tathagata Das <[email protected]>
Date: 2015-12-10T04:47:15Z
[SPARK-12244][SPARK-12245][STREAMING] Rename trackStateByKey to
mapWithState and change tracking function signature
SPARK-12244:
Based on feedback from early users and personal experience attempting to
explain it, the name trackStateByKey had two problem.
"trackState" is a completely new term which really does not give any
intuition on what the operation is
the resultant data stream of objects returned by the function is called in
docs as the "emitted" data for the lack of a better.
"mapWithState" makes sense because the API is like a mapping function like
(Key, Value) => T with State as an additional parameter. The resultant data
stream is "mapped data". So both problems are solved.
SPARK-12245:
From initial experiences, not having the key in the function makes it hard
to return mapped stuff, as the whole information of the records is not there.
Basically the user is restricted to doing something like mapValue() instead of
map(). So adding the key as a parameter.
Author: Tathagata Das <[email protected]>
Closes #10224 from tdas/rename.
commit b5e5812f9ef8aa8d133a75bb8aa8dd8680130efa
Author: bomeng <[email protected]>
Date: 2015-12-10T12:53:53Z
[SPARK-12136][STREAMING] rddToFileName does not properly handle prefix and
suffix parameters
The original code does not properly handle the cases where the prefix is
null, but suffix is not null - the suffix should be used but is not.
The fix is using StringBuilder to construct the proper file name.
Author: bomeng <[email protected]>
Author: Bo Meng <[email protected]>
Closes #10185 from bomeng/SPARK-12136.
(cherry picked from commit e29704f90dfe67d9e276d242699ac0a00f64fb91)
Signed-off-by: Sean Owen <[email protected]>
commit f939c71b187cff3a5bb63aa3659429b6efb0626d
Author: Reynold Xin <[email protected]>
Date: 2015-12-10T14:23:10Z
[SPARK-12242][SQL] Add DataFrame.transform method
Author: Reynold Xin <[email protected]>
Closes #10226 from rxin/df-transform.
(cherry picked from commit 76540b6df5370b463277d3498097b2cc2d2e97a8)
Signed-off-by: Reynold Xin <[email protected]>
commit b7b9f772751dc4ea7eb28a2bdb897a04e563fafa
Author: Yanbo Liang <[email protected]>
Date: 2015-12-10T17:44:53Z
[SPARK-12198][SPARKR] SparkR support read.parquet and deprecate parquetFile
SparkR support ```read.parquet``` and deprecate ```parquetFile```. This
change is similar with #10145 for ```jsonFile```.
Author: Yanbo Liang <[email protected]>
Closes #10191 from yanboliang/spark-12198.
(cherry picked from commit eeb58722ad73441eeb5f35f864be3c5392cfd426)
Signed-off-by: Shivaram Venkataraman <[email protected]>
commit e65c88536ad1843a45e0fe3cf1edadfdf4ad3460
Author: Yuhao Yang <[email protected]>
Date: 2015-12-10T18:15:50Z
[SPARK-11602][MLLIB] Refine visibility for 1.6 scala API audit
jira: https://issues.apache.org/jira/browse/SPARK-11602
Made a pass on the API change of 1.6. Open the PR for efficient discussion.
Author: Yuhao Yang <[email protected]>
Closes #9939 from hhbyyh/auditScala.
(cherry picked from commit 9fba9c8004d2b97549e5456fa7918965bec27336)
Signed-off-by: Joseph K. Bradley <[email protected]>
commit 93ef2463820928f434f9fb1542bc30cfb1cec9aa
Author: Yanbo Liang <[email protected]>
Date: 2015-12-10T18:18:58Z
[SPARK-12234][SPARKR] Fix ```subset``` function error when only set
```select``` argument
Fix ```subset``` function error when only set ```select``` argument. Please
refer to the [JIRA](https://issues.apache.org/jira/browse/SPARK-12234) about
the error and how to reproduce it.
cc sun-rui felixcheung shivaram
Author: Yanbo Liang <[email protected]>
Closes #10217 from yanboliang/spark-12234.
(cherry picked from commit d9d354ed40eec56b3f03d32f4e2629d367b1bf02)
Signed-off-by: Shivaram Venkataraman <[email protected]>
commit e541f703d72d3dd3ad96db55650c5b1a1a5a38e2
Author: Cheng Lian <[email protected]>
Date: 2015-12-10T18:19:44Z
[SPARK-12012][SQL][BRANCH-1.6] Show more comprehensive PhysicalRDD metadata
when visualizing SQL query plan
This PR backports PR #10004 to branch-1.6
It adds a private[sql] method metadata to SparkPlan, which can be used to
describe detail information about a physical plan during visualization.
Specifically, this PR uses this method to provide details of PhysicalRDDs
translated from a data source relation.
Author: Cheng Lian <[email protected]>
Closes #10250 from liancheng/spark-12012.for-1.6.
commit 594fafc6122ad9c6b24bdb4a434d97158c7745f3
Author: Yin Huai <[email protected]>
Date: 2015-12-10T20:03:29Z
[SPARK-12250][SQL] Allow users to define a UDAF without providing details
of its inputSchema
https://issues.apache.org/jira/browse/SPARK-12250
Author: Yin Huai <[email protected]>
Closes #10236 from yhuai/SPARK-12250.
(cherry picked from commit bc5f56aa60a430244ffa0cacd81c0b1ecbf8d68f)
Signed-off-by: Yin Huai <[email protected]>
commit d0307deaa29d5fcf1c675f9367c26aa6a3db3fba
Author: Timothy Hunter <[email protected]>
Date: 2015-12-10T20:50:46Z
[SPARK-12212][ML][DOC] Clarifies the difference between spark.ml,
spark.mllib and mllib in the documentation.
Replaces a number of occurences of `MLlib` in the documentation that were
meant to refer to the `spark.mllib` package instead. It should clarify for new
users the difference between `spark.mllib` (the package) and MLlib (the
umbrella project for ML in spark).
It also removes some files that I forgot to delete with #10207
Author: Timothy Hunter <[email protected]>
Closes #10234 from thunterdb/12212.
(cherry picked from commit 2ecbe02d5b28ee562d10c1735244b90a08532c9e)
Signed-off-by: Joseph K. Bradley <[email protected]>
commit 9870e5c7af87190167ca3845ede918671b9420ca
Author: Josh Rosen <[email protected]>
Date: 2015-12-10T23:29:04Z
[SPARK-12251] Document and improve off-heap memory configurations
This patch adds documentation for Spark configurations that affect off-heap
memory and makes some naming and validation improvements for those configs.
- Change `spark.memory.offHeapSize` to `spark.memory.offHeap.size`. This is
fine because this configuration has not shipped in any Spark release yet (it's
new in Spark 1.6).
- Deprecated `spark.unsafe.offHeap` in favor of a new
`spark.memory.offHeap.enabled` configuration. The motivation behind this change
is to gather all memory-related configurations under the same prefix.
- Add a check which prevents users from setting
`spark.memory.offHeap.enabled=true` when `spark.memory.offHeap.size == 0`.
After SPARK-11389 (#9344), which was committed in Spark 1.6, Spark enforces a
hard limit on the amount of off-heap memory that it will allocate to tasks. As
a result, enabling off-heap execution memory without setting
`spark.memory.offHeap.size` will lead to immediate OOMs. The new configuration
validation makes this scenario easier to diagnose, helping to avoid user
confusion.
- Document these configurations on the configuration page.
Author: Josh Rosen <[email protected]>
Closes #10237 from JoshRosen/SPARK-12251.
(cherry picked from commit 23a9e62bad9669e9ff5dc4bd714f58d12f9be0b5)
Signed-off-by: Andrew Or <[email protected]>
commit c247b6a6546e12e3c6992c40cad1881d56aefd6f
Author: Andrew Or <[email protected]>
Date: 2015-12-10T23:30:08Z
[SPARK-12155][SPARK-12253] Fix executor OOM in unified memory management
**Problem.** In unified memory management, acquiring execution memory may
lead to eviction of storage memory. However, the space freed from evicting
cached blocks is distributed among all active tasks. Thus, an incorrect upper
bound on the execution memory per task can cause the acquisition to fail,
leading to OOM's and premature spills.
**Example.** Suppose total memory is 1000B, cached blocks occupy 900B,
`spark.memory.storageFraction` is 0.4, and there are two active tasks. In this
case, the cap on task execution memory is 100B / 2 = 50B. If task A tries to
acquire 200B, it will evict 100B of storage but can only acquire 50B because of
the incorrect cap. For another example, see this [regression
test](https://github.com/andrewor14/spark/blob/fix-oom/core/src/test/scala/org/apache/spark/memory/UnifiedMemoryManagerSuite.scala#L233)
that I stole from JoshRosen.
**Solution.** Fix the cap on task execution memory. It should take into
account the space that could have been freed by storage in addition to the
current amount of memory available to execution. In the example above, the
correct cap should have been 600B / 2 = 300B.
This patch also guards against the race condition (SPARK-12253):
(1) Existing tasks collectively occupy all execution memory
(2) New task comes in and blocks while existing tasks spill
(3) After tasks finish spilling, another task jumps in and puts in a large
block, stealing the freed memory
(4) New task still cannot acquire memory and goes back to sleep
Author: Andrew Or <[email protected]>
Closes #10240 from andrewor14/fix-oom.
(cherry picked from commit 5030923ea8bb94ac8fa8e432de9fc7089aa93986)
Signed-off-by: Andrew Or <[email protected]>
commit 5d3722f8e5cdb4abd946ea18950225919af53a11
Author: jerryshao <[email protected]>
Date: 2015-12-10T23:31:46Z
[STREAMING][DOC][MINOR] Update the description of direct Kafka stream doc
With the merge of
[SPARK-8337](https://issues.apache.org/jira/browse/SPARK-8337), now the Python
API has the same functionalities compared to Scala/Java, so here changing the
description to make it more precise.
zsxwing tdas , please review, thanks a lot.
Author: jerryshao <[email protected]>
Closes #10246 from jerryshao/direct-kafka-doc-update.
(cherry picked from commit 24d3357d66e14388faf8709b368edca70ea96432)
Signed-off-by: Shixiong Zhu <[email protected]>
commit d09af2cb4237cca9ac72aacb9abb822a2982a820
Author: Davies Liu <[email protected]>
Date: 2015-12-11T01:22:18Z
[SPARK-12258][SQL] passing null into ScalaUDF
Check nullability and passing them into ScalaUDF.
Closes #10249
Author: Davies Liu <[email protected]>
Closes #10259 from davies/udf_null.
(cherry picked from commit b1b4ee7f3541d92c8bc2b0b4fdadf46cfdb09504)
Signed-off-by: Yin Huai <[email protected]>
commit 3e39925f9296bc126adf3f6828a0adf306900c0a
Author: Patrick Wendell <[email protected]>
Date: 2015-12-11T02:45:36Z
Preparing Spark release v1.6.0-rc2
commit 250249e26466ff0d6ee6f8ae34f0225285c9bb9b
Author: Patrick Wendell <[email protected]>
Date: 2015-12-11T02:45:42Z
Preparing development version 1.6.0-SNAPSHOT
commit eec36607f9fc92b6c4d306e3930fcf03961625eb
Author: Davies Liu <[email protected]>
Date: 2015-12-11T19:15:53Z
[SPARK-12258] [SQL] passing null into ScalaUDF (follow-up)
This is a follow-up PR for #10259
Author: Davies Liu <[email protected]>
Closes #10266 from davies/null_udf2.
(cherry picked from commit c119a34d1e9e599e302acfda92e5de681086a19f)
Signed-off-by: Davies Liu <[email protected]>
commit 23f8dfd45187cb8f2216328ab907ddb5fbdffd0b
Author: Patrick Wendell <[email protected]>
Date: 2015-12-11T19:25:03Z
Preparing Spark release v1.6.0-rc2
commit 2e4523161ddf2417f2570bb75cc2d6694813adf5
Author: Patrick Wendell <[email protected]>
Date: 2015-12-11T19:25:09Z
Preparing development version 1.6.0-SNAPSHOT
----
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enabled but not working, please
contact infrastructure at [email protected] or file a JIRA ticket
with INFRA.
---
---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]