spark git commit: [SPARK-10128] [STREAMING] Used correct classloader to deserialize WAL data

2015-08-19 Thread tdas
Repository: spark Updated Branches: refs/heads/branch-1.5 63922fa4d -> 71aa54755 [SPARK-10128] [STREAMING] Used correct classloader to deserialize WAL data Recovering Kinesis sequence numbers from WAL leads to classnotfoundexception because the ObjectInputStream does not use the correct class

spark git commit: [SPARK-10128] [STREAMING] Used correct classloader to deserialize WAL data

2015-08-19 Thread tdas
Repository: spark Updated Branches: refs/heads/master 73431d8af -> b762f9920 [SPARK-10128] [STREAMING] Used correct classloader to deserialize WAL data Recovering Kinesis sequence numbers from WAL leads to classnotfoundexception because the ObjectInputStream does not use the correct classload

spark git commit: [SPARK-10125] [STREAMING] Fix a potential deadlock in JobGenerator.stop

2015-08-19 Thread tdas
Repository: spark Updated Branches: refs/heads/branch-1.5 a3ed2c31e -> 63922fa4d [SPARK-10125] [STREAMING] Fix a potential deadlock in JobGenerator.stop Because `lazy val` uses `this` lock, if JobGenerator.stop and JobGenerator.doCheckpoint (JobGenerator.shouldCheckpoint has not yet been ini

spark git commit: [SPARK-10124] [MESOS] Fix removing queued driver in mesos cluster mode.

2015-08-19 Thread andrewor14
Repository: spark Updated Branches: refs/heads/branch-1.5 16414dae0 -> a3ed2c31e [SPARK-10124] [MESOS] Fix removing queued driver in mesos cluster mode. Currently the spark applications can be queued to the Mesos cluster dispatcher, but when multiple jobs are in queue we don't handle removing

spark git commit: [SPARK-10124] [MESOS] Fix removing queued driver in mesos cluster mode.

2015-08-19 Thread andrewor14
Repository: spark Updated Branches: refs/heads/master affc8a887 -> 73431d8af [SPARK-10124] [MESOS] Fix removing queued driver in mesos cluster mode. Currently the spark applications can be queued to the Mesos cluster dispatcher, but when multiple jobs are in queue we don't handle removing job

spark git commit: [SPARK-10125] [STREAMING] Fix a potential deadlock in JobGenerator.stop

2015-08-19 Thread tdas
Repository: spark Updated Branches: refs/heads/master 1f29d502e -> affc8a887 [SPARK-10125] [STREAMING] Fix a potential deadlock in JobGenerator.stop Because `lazy val` uses `this` lock, if JobGenerator.stop and JobGenerator.doCheckpoint (JobGenerator.shouldCheckpoint has not yet been initial

spark git commit: [SPARK-9812] [STREAMING] Fix Python 3 compatibility issue in PySpark Streaming and some docs

2015-08-19 Thread tdas
Repository: spark Updated Branches: refs/heads/branch-1.5 321cb99ca -> 16414dae0 [SPARK-9812] [STREAMING] Fix Python 3 compatibility issue in PySpark Streaming and some docs This PR includes the following fixes: 1. Use `range` instead of `xrange` in `queue_stream.py` to support Python 3. 2. F

spark git commit: [SPARK-9812] [STREAMING] Fix Python 3 compatibility issue in PySpark Streaming and some docs

2015-08-19 Thread tdas
Repository: spark Updated Branches: refs/heads/master 2f2686a73 -> 1f29d502e [SPARK-9812] [STREAMING] Fix Python 3 compatibility issue in PySpark Streaming and some docs This PR includes the following fixes: 1. Use `range` instead of `xrange` in `queue_stream.py` to support Python 3. 2. Fix t

[5/5] spark git commit: [SPARK-9242] [SQL] Audit UDAF interface.

2015-08-19 Thread rxin
[SPARK-9242] [SQL] Audit UDAF interface. A few minor changes: 1. Improved documentation 2. Rename apply(distinct) to distinct. 3. Changed MutableAggregationBuffer from a trait to an abstract class. 4. Renamed returnDataType to dataType to be more consistent with other expressions. And unrel

[4/5] spark git commit: [SPARK-9242] [SQL] Audit UDAF interface.

2015-08-19 Thread rxin
http://git-wip-us.apache.org/repos/asf/spark/blob/2f2686a7/sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/expressions/datetimeFunctions.scala -- diff --git a/sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/exp

[5/5] spark git commit: [SPARK-9242] [SQL] Audit UDAF interface.

2015-08-19 Thread rxin
[SPARK-9242] [SQL] Audit UDAF interface. A few minor changes: 1. Improved documentation 2. Rename apply(distinct) to distinct. 3. Changed MutableAggregationBuffer from a trait to an abstract class. 4. Renamed returnDataType to dataType to be more consistent with other expressions. And unrel

[3/5] spark git commit: [SPARK-9242] [SQL] Audit UDAF interface.

2015-08-19 Thread rxin
http://git-wip-us.apache.org/repos/asf/spark/blob/2f2686a7/sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/expressions/math.scala -- diff --git a/sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/expressions/math

[3/5] spark git commit: [SPARK-9242] [SQL] Audit UDAF interface.

2015-08-19 Thread rxin
http://git-wip-us.apache.org/repos/asf/spark/blob/321cb99c/sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/expressions/math.scala -- diff --git a/sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/expressions/math

[1/5] spark git commit: [SPARK-9242] [SQL] Audit UDAF interface.

2015-08-19 Thread rxin
Repository: spark Updated Branches: refs/heads/branch-1.5 56a37b01f -> 321cb99ca http://git-wip-us.apache.org/repos/asf/spark/blob/321cb99c/sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/expressions/stringOperations.scala --

[2/5] spark git commit: [SPARK-9242] [SQL] Audit UDAF interface.

2015-08-19 Thread rxin
http://git-wip-us.apache.org/repos/asf/spark/blob/321cb99c/sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/expressions/regexpExpressions.scala -- diff --git a/sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/exp

[2/5] spark git commit: [SPARK-9242] [SQL] Audit UDAF interface.

2015-08-19 Thread rxin
http://git-wip-us.apache.org/repos/asf/spark/blob/2f2686a7/sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/expressions/regexpExpressions.scala -- diff --git a/sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/exp

[1/5] spark git commit: [SPARK-9242] [SQL] Audit UDAF interface.

2015-08-19 Thread rxin
Repository: spark Updated Branches: refs/heads/master ba5f7e184 -> 2f2686a73 http://git-wip-us.apache.org/repos/asf/spark/blob/2f2686a7/sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/expressions/stringOperations.scala --

[4/5] spark git commit: [SPARK-9242] [SQL] Audit UDAF interface.

2015-08-19 Thread rxin
http://git-wip-us.apache.org/repos/asf/spark/blob/321cb99c/sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/expressions/datetimeFunctions.scala -- diff --git a/sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/exp

spark git commit: [SPARK-10035] [SQL] Parquet filters does not process EqualNullSafe filter.

2015-08-19 Thread lian
Repository: spark Updated Branches: refs/heads/master 8e0a072f7 -> ba5f7e184 [SPARK-10035] [SQL] Parquet filters does not process EqualNullSafe filter. As I talked with Lian, 1. I added EquelNullSafe to ParquetFilters - It uses the same equality comparison filter with EqualTo since the Parqu

spark git commit: [SPARK-9895] User Guide for RFormula Feature Transformer

2015-08-19 Thread meng
Repository: spark Updated Branches: refs/heads/master b0dbaec4f -> 8e0a072f7 [SPARK-9895] User Guide for RFormula Feature Transformer mengxr Author: Eric Liang Closes #8293 from ericl/docs-2. Project: http://git-wip-us.apache.org/repos/asf/spark/repo Commit: http://git-wip-us.apache.org/r

spark git commit: [SPARK-9895] User Guide for RFormula Feature Transformer

2015-08-19 Thread meng
Repository: spark Updated Branches: refs/heads/branch-1.5 5c749c82c -> 56a37b01f [SPARK-9895] User Guide for RFormula Feature Transformer mengxr Author: Eric Liang Closes #8293 from ericl/docs-2. (cherry picked from commit 8e0a072f78b4902d5f7ccc6b15232ed202a117f9) Signed-off-by: Xiangrui M

spark git commit: [SPARK-6489] [SQL] add column pruning for Generate

2015-08-19 Thread marmbrus
Repository: spark Updated Branches: refs/heads/branch-1.5 a59475f5b -> 5c749c82c [SPARK-6489] [SQL] add column pruning for Generate This PR takes over https://github.com/apache/spark/pull/5358 Author: Wenchen Fan Closes #8268 from cloud-fan/6489. (cherry picked from commit b0dbaec4f942a47a

spark git commit: [SPARK-6489] [SQL] add column pruning for Generate

2015-08-19 Thread marmbrus
Repository: spark Updated Branches: refs/heads/master e0dd1309a -> b0dbaec4f [SPARK-6489] [SQL] add column pruning for Generate This PR takes over https://github.com/apache/spark/pull/5358 Author: Wenchen Fan Closes #8268 from cloud-fan/6489. Project: http://git-wip-us.apache.org/repos/as

spark git commit: [SPARK-10119] [CORE] Fix isDynamicAllocationEnabled when config is expliticly disabled.

2015-08-19 Thread vanzin
Repository: spark Updated Branches: refs/heads/master 373a376c0 -> e0dd1309a [SPARK-10119] [CORE] Fix isDynamicAllocationEnabled when config is expliticly disabled. Author: Marcelo Vanzin Closes #8316 from vanzin/SPARK-10119. Project: http://git-wip-us.apache.org/repos/asf/spark/repo Comm

spark git commit: [SPARK-10119] [CORE] Fix isDynamicAllocationEnabled when config is expliticly disabled.

2015-08-19 Thread vanzin
Repository: spark Updated Branches: refs/heads/branch-1.5 1494d58d2 -> a59475f5b [SPARK-10119] [CORE] Fix isDynamicAllocationEnabled when config is expliticly disabled. Author: Marcelo Vanzin Closes #8316 from vanzin/SPARK-10119. (cherry picked from commit e0dd1309ac248375f429639801923570f

spark git commit: [SPARK-10083] [SQL] CaseWhen should support type coercion of DecimalType and FractionalType

2015-08-19 Thread marmbrus
Repository: spark Updated Branches: refs/heads/branch-1.5 b32a31d64 -> 1494d58d2 [SPARK-10083] [SQL] CaseWhen should support type coercion of DecimalType and FractionalType create t1 (a decimal(7, 2), b long); select case when 1=1 then a else 1.0 end from t1; select case when 1=1 then a else

spark git commit: [SPARK-10083] [SQL] CaseWhen should support type coercion of DecimalType and FractionalType

2015-08-19 Thread marmbrus
Repository: spark Updated Branches: refs/heads/master f3ff4c41d -> 373a376c0 [SPARK-10083] [SQL] CaseWhen should support type coercion of DecimalType and FractionalType create t1 (a decimal(7, 2), b long); select case when 1=1 then a else 1.0 end from t1; select case when 1=1 then a else b en

spark git commit: [SPARK-9899] [SQL] Disables customized output committer when speculation is on

2015-08-19 Thread marmbrus
Repository: spark Updated Branches: refs/heads/branch-1.5 d9dfd43d4 -> b32a31d64 [SPARK-9899] [SQL] Disables customized output committer when speculation is on Speculation hates direct output committer, as there are multiple corner cases that may cause data corruption and/or data loss. Pleas

spark git commit: [SPARK-9899] [SQL] Disables customized output committer when speculation is on

2015-08-19 Thread marmbrus
Repository: spark Updated Branches: refs/heads/master 1f4c4fe6d -> f3ff4c41d [SPARK-9899] [SQL] Disables customized output committer when speculation is on Speculation hates direct output committer, as there are multiple corner cases that may cause data corruption and/or data loss. Please se

spark git commit: [SPARK-10090] [SQL] fix decimal scale of division

2015-08-19 Thread marmbrus
Repository: spark Updated Branches: refs/heads/branch-1.5 77269fcb5 -> d9dfd43d4 [SPARK-10090] [SQL] fix decimal scale of division We should rounding the result of multiply/division of decimal to expected precision/scale, also check overflow. Author: Davies Liu Closes #8287 from davies/dec

spark git commit: [SPARK-10090] [SQL] fix decimal scale of division

2015-08-19 Thread marmbrus
Repository: spark Updated Branches: refs/heads/master 21bdbe9fe -> 1f4c4fe6d [SPARK-10090] [SQL] fix decimal scale of division We should rounding the result of multiply/division of decimal to expected precision/scale, also check overflow. Author: Davies Liu Closes #8287 from davies/decimal

spark git commit: [SPARK-9627] [SQL] Stops using Scala runtime reflection in DictionaryEncoding

2015-08-19 Thread marmbrus
Repository: spark Updated Branches: refs/heads/branch-1.5 afaed7ef4 -> 77269fcb5 [SPARK-9627] [SQL] Stops using Scala runtime reflection in DictionaryEncoding `DictionaryEncoding` uses Scala runtime reflection to avoid boxing costs while building the directory array. However, this code path m

spark git commit: [SPARK-9627] [SQL] Stops using Scala runtime reflection in DictionaryEncoding

2015-08-19 Thread marmbrus
Repository: spark Updated Branches: refs/heads/master 08887369c -> 21bdbe9fe [SPARK-9627] [SQL] Stops using Scala runtime reflection in DictionaryEncoding `DictionaryEncoding` uses Scala runtime reflection to avoid boxing costs while building the directory array. However, this code path may h

spark git commit: [SPARK-10073] [SQL] Python withColumn should replace the old column

2015-08-19 Thread marmbrus
Repository: spark Updated Branches: refs/heads/branch-1.5 829c33a4b -> afaed7ef4 [SPARK-10073] [SQL] Python withColumn should replace the old column DataFrame.withColumn in Python should be consistent with the Scala one (replacing the existing column that has the same name). cc marmbrus Au

spark git commit: [SPARK-10073] [SQL] Python withColumn should replace the old column

2015-08-19 Thread marmbrus
Repository: spark Updated Branches: refs/heads/master e05da5cb5 -> 08887369c [SPARK-10073] [SQL] Python withColumn should replace the old column DataFrame.withColumn in Python should be consistent with the Scala one (replacing the existing column that has the same name). cc marmbrus Author

spark git commit: [SPARK-10107] [SQL] fix NPE in format_number

2015-08-19 Thread rxin
Repository: spark Updated Branches: refs/heads/branch-1.5 8c0a5a248 -> 1038f677b [SPARK-10107] [SQL] fix NPE in format_number Author: Davies Liu Closes #8305 from davies/format_number. (cherry picked from commit e05da5cb5ea253e6372f648fc8203204f2a8df8d) Signed-off-by: Reynold Xin Project

spark git commit: [SPARK-10087] [CORE] [BRANCH-1.5] Disable spark.shuffle.reduceLocality.enabled by default.

2015-08-19 Thread rxin
Repository: spark Updated Branches: refs/heads/branch-1.5 1038f677b -> 829c33a4b [SPARK-10087] [CORE] [BRANCH-1.5] Disable spark.shuffle.reduceLocality.enabled by default. https://issues.apache.org/jira/browse/SPARK-10087 In some cases, when spark.shuffle.reduceLocality.enabled is enabled, w

spark git commit: [SPARK-10107] [SQL] fix NPE in format_number

2015-08-19 Thread rxin
Repository: spark Updated Branches: refs/heads/master f3391ff2b -> e05da5cb5 [SPARK-10107] [SQL] fix NPE in format_number Author: Davies Liu Closes #8305 from davies/format_number. Project: http://git-wip-us.apache.org/repos/asf/spark/repo Commit: http://git-wip-us.apache.org/repos/asf/spa

spark git commit: [SPARK-8889] [CORE] Fix for OOM for graph creation

2015-08-19 Thread srowen
Repository: spark Updated Branches: refs/heads/master 5b62bef8c -> f3391ff2b [SPARK-8889] [CORE] Fix for OOM for graph creation Fix for OOM for graph creation Author: Joshi Author: Rekha Joshi Closes #7602 from rekhajoshm/SPARK-8889. Project: http://git-wip-us.apache.org/repos/asf/spark/

spark git commit: [SPARK-8918] [MLLIB] [DOC] Add @since tags to mllib.clustering

2015-08-19 Thread meng
Repository: spark Updated Branches: refs/heads/branch-1.5 ba369258d -> 8c0a5a248 [SPARK-8918] [MLLIB] [DOC] Add @since tags to mllib.clustering This continues the work from #8256. I removed `since` tags from private/protected/local methods/variables (see https://github.com/apache/spark/commi

spark git commit: [SPARK-8918] [MLLIB] [DOC] Add @since tags to mllib.clustering

2015-08-19 Thread meng
Repository: spark Updated Branches: refs/heads/master d898c33f7 -> 5b62bef8c [SPARK-8918] [MLLIB] [DOC] Add @since tags to mllib.clustering This continues the work from #8256. I removed `since` tags from private/protected/local methods/variables (see https://github.com/apache/spark/commit/72

spark git commit: [SPARK-10106] [SPARKR] Add `ifelse` Column function to SparkR

2015-08-19 Thread shivaram
Repository: spark Updated Branches: refs/heads/master 28a98464e -> d898c33f7 [SPARK-10106] [SPARKR] Add `ifelse` Column function to SparkR ### JIRA [[SPARK-10106] Add `ifelse` Column function to SparkR - ASF JIRA](https://issues.apache.org/jira/browse/SPARK-10106) Author: Yu ISHIKAWA Close

spark git commit: [SPARK-10106] [SPARKR] Add `ifelse` Column function to SparkR

2015-08-19 Thread shivaram
Repository: spark Updated Branches: refs/heads/branch-1.5 f25c32475 -> ba369258d [SPARK-10106] [SPARKR] Add `ifelse` Column function to SparkR ### JIRA [[SPARK-10106] Add `ifelse` Column function to SparkR - ASF JIRA](https://issues.apache.org/jira/browse/SPARK-10106) Author: Yu ISHIKAWA C

svn commit: r1696648 - in /spark: mllib/index.md site/mllib/index.html

2015-08-19 Thread meng
Author: meng Date: Wed Aug 19 19:11:08 2015 New Revision: 1696648 URL: http://svn.apache.org/r1696648 Log: update MLlib page for 1.5 Modified: spark/mllib/index.md spark/site/mllib/index.html Modified: spark/mllib/index.md URL: http://svn.apache.org/viewvc/spark/mllib/index.md?rev=16966

spark git commit: [SPARK-10097] Adds `shouldMaximize` flag to `ml.evaluation.Evaluator`

2015-08-19 Thread jkbradley
Repository: spark Updated Branches: refs/heads/branch-1.5 a8e880818 -> f25c32475 [SPARK-10097] Adds `shouldMaximize` flag to `ml.evaluation.Evaluator` Previously, users of evaluator (`CrossValidator` and `TrainValidationSplit`) would only maximize the metric in evaluator, leading to a hacky s

spark git commit: [SPARK-10097] Adds `shouldMaximize` flag to `ml.evaluation.Evaluator`

2015-08-19 Thread jkbradley
Repository: spark Updated Branches: refs/heads/master 5fd53c64b -> 28a98464e [SPARK-10097] Adds `shouldMaximize` flag to `ml.evaluation.Evaluator` Previously, users of evaluator (`CrossValidator` and `TrainValidationSplit`) would only maximize the metric in evaluator, leading to a hacky solut

spark git commit: [SPARK-9833] [YARN] Add options to disable delegation token retrieval.

2015-08-19 Thread vanzin
Repository: spark Updated Branches: refs/heads/master 2fcb9cb95 -> 5fd53c64b [SPARK-9833] [YARN] Add options to disable delegation token retrieval. This allows skipping the code that tries to talk to Hive and HBase to fetch delegation tokens, in case that somehow conflicts with the application

spark git commit: [SPARK-9856] [SPARKR] Add expression functions into SparkR whose params are complicated

2015-08-19 Thread shivaram
Repository: spark Updated Branches: refs/heads/branch-1.5 bebe63dfe -> a8e880818 [SPARK-9856] [SPARKR] Add expression functions into SparkR whose params are complicated I added lots of Column functinos into SparkR. And I also added `rand(seed: Int)` and `randn(seed: Int)` in Scala. Since we

spark git commit: [SPARK-9856] [SPARKR] Add expression functions into SparkR whose params are complicated

2015-08-19 Thread shivaram
Repository: spark Updated Branches: refs/heads/master f3e177917 -> 2fcb9cb95 [SPARK-9856] [SPARKR] Add expression functions into SparkR whose params are complicated I added lots of Column functinos into SparkR. And I also added `rand(seed: Int)` and `randn(seed: Int)` in Scala. Since we need

spark git commit: [SPARK-5754] [YARN] Spark/Yarn/Windows driver/executor escaping Fix

2015-08-19 Thread vanzin
Repository: spark Updated Branches: refs/heads/master 802b5b879 -> f3e177917 [SPARK-5754] [YARN] Spark/Yarn/Windows driver/executor escaping Fix This is my retry to suggest a fix for using Spark on Yarn on Windows. The former request lacked coding style which I hope to have learned to do bett

spark git commit: [SPARK-10084] [MLLIB] [DOC] Add Python example for mllib FP-growth user guide

2015-08-19 Thread meng
Repository: spark Updated Branches: refs/heads/branch-1.5 f8dc42770 -> bebe63dfe [SPARK-10084] [MLLIB] [DOC] Add Python example for mllib FP-growth user guide 1, Add Python example for mllib FP-growth user guide. 2, Correct mistakes of Scala and Java examples. Author: Yanbo Liang Closes #82

spark git commit: [SPARK-10084] [MLLIB] [DOC] Add Python example for mllib FP-growth user guide

2015-08-19 Thread meng
Repository: spark Updated Branches: refs/heads/master 39e4ebd52 -> 802b5b879 [SPARK-10084] [MLLIB] [DOC] Add Python example for mllib FP-growth user guide 1, Add Python example for mllib FP-growth user guide. 2, Correct mistakes of Scala and Java examples. Author: Yanbo Liang Closes #8279 f

spark git commit: [SPARK-10060] [ML] [DOC] spark.ml DecisionTree user guide

2015-08-19 Thread meng
Repository: spark Updated Branches: refs/heads/branch-1.5 522b0b646 -> f8dc42770 [SPARK-10060] [ML] [DOC] spark.ml DecisionTree user guide New user guide section ml-decision-tree.md, including code examples. I have run all examples, including the Java ones. CC: manishamde yanboliang mengxr

spark git commit: [SPARK-10060] [ML] [DOC] spark.ml DecisionTree user guide

2015-08-19 Thread meng
Repository: spark Updated Branches: refs/heads/master 3d16a5450 -> 39e4ebd52 [SPARK-10060] [ML] [DOC] spark.ml DecisionTree user guide New user guide section ml-decision-tree.md, including code examples. I have run all examples, including the Java ones. CC: manishamde yanboliang mengxr Auth

spark git commit: [SPARK-8949] Print warnings when using preferred locations feature

2015-08-19 Thread srowen
Repository: spark Updated Branches: refs/heads/branch-1.5 5553f02be -> 522b0b646 [SPARK-8949] Print warnings when using preferred locations feature Add warnings according to SPARK-8949 in `SparkContext` - warnings in scaladoc - log warnings when preferred locations feature is used through

spark git commit: [SPARK-8949] Print warnings when using preferred locations feature

2015-08-19 Thread srowen
Repository: spark Updated Branches: refs/heads/master ba2a07e2b -> 3d16a5450 [SPARK-8949] Print warnings when using preferred locations feature Add warnings according to SPARK-8949 in `SparkContext` - warnings in scaladoc - log warnings when preferred locations feature is used through `Sp

spark git commit: [SPARK-9977] [DOCS] Update documentation for StringIndexer

2015-08-19 Thread srowen
Repository: spark Updated Branches: refs/heads/branch-1.5 e56bcc638 -> 5553f02be [SPARK-9977] [DOCS] Update documentation for StringIndexer By using `StringIndexer`, we can obtain indexed label on new column. So a following estimator should use this new column through pipeline if it wants to

spark git commit: [SPARK-9977] [DOCS] Update documentation for StringIndexer

2015-08-19 Thread srowen
Repository: spark Updated Branches: refs/heads/master 865a3df3d -> ba2a07e2b [SPARK-9977] [DOCS] Update documentation for StringIndexer By using `StringIndexer`, we can obtain indexed label on new column. So a following estimator should use this new column through pipeline if it wants to use

spark git commit: [DOCS] [SQL] [PYSPARK] Fix typo in ntile function

2015-08-19 Thread srowen
Repository: spark Updated Branches: refs/heads/branch-1.5 561390dbc -> e56bcc638 [DOCS] [SQL] [PYSPARK] Fix typo in ntile function Fix typo in ntile function. Author: Moussa Taifi Closes #8261 from moutai/patch-2. (cherry picked from commit 865a3df3d578c0442c97d749c81f554b560da406) Signed-

spark git commit: [DOCS] [SQL] [PYSPARK] Fix typo in ntile function

2015-08-19 Thread srowen
Repository: spark Updated Branches: refs/heads/master f141efeaf -> 865a3df3d [DOCS] [SQL] [PYSPARK] Fix typo in ntile function Fix typo in ntile function. Author: Moussa Taifi Closes #8261 from moutai/patch-2. Project: http://git-wip-us.apache.org/repos/asf/spark/repo Commit: http://git-w

spark git commit: [SPARK-10070] [DOCS] Remove Guava dependencies in user guides

2015-08-19 Thread srowen
Repository: spark Updated Branches: refs/heads/master b23c4d3ff -> f141efeaf [SPARK-10070] [DOCS] Remove Guava dependencies in user guides `Lists.newArrayList` -> `Arrays.asList` CC jkbradley feynmanliang Anybody into replacing usages of `Lists.newArrayList` in the examples / source code to

spark git commit: [SPARK-10070] [DOCS] Remove Guava dependencies in user guides

2015-08-19 Thread srowen
Repository: spark Updated Branches: refs/heads/branch-1.5 417852f7c -> 561390dbc [SPARK-10070] [DOCS] Remove Guava dependencies in user guides `Lists.newArrayList` -> `Arrays.asList` CC jkbradley feynmanliang Anybody into replacing usages of `Lists.newArrayList` in the examples / source cod

spark git commit: Fix Broken Link

2015-08-19 Thread rxin
Repository: spark Updated Branches: refs/heads/branch-1.5 392bd19d6 -> 417852f7c Fix Broken Link Link was broken because it included tick marks. Author: Bill Chambers Closes #8302 from anabranch/patch-1. (cherry picked from commit b23c4d3ffc36e47c057360c611d8ab1a13877699) Signed-off-by: Re

spark git commit: Fix Broken Link

2015-08-19 Thread rxin
Repository: spark Updated Branches: refs/heads/master bc9a0e032 -> b23c4d3ff Fix Broken Link Link was broken because it included tick marks. Author: Bill Chambers Closes #8302 from anabranch/patch-1. Project: http://git-wip-us.apache.org/repos/asf/spark/repo Commit: http://git-wip-us.apac