Repository: spark
Updated Branches:
refs/heads/master 9d65f2712 -> 3e2864e40
[SPARK-3139] Made ContextCleaner to not block on shuffles
As a workaround for SPARK-3015, the ContextCleaner was made "blocking", that
is, it cleaned items one-by-one. But shuffles can take a long time to be
deleted
Repository: spark
Updated Branches:
refs/heads/branch-1.1 6f82a4b13 -> 5cf1e4401
[SPARK-3139] Made ContextCleaner to not block on shuffles
As a workaround for SPARK-3015, the ContextCleaner was made "blocking", that
is, it cleaned items one-by-one. But shuffles can take a long time to be
del
Repository: spark
Updated Branches:
refs/heads/master 3e2864e40 -> e1139dd60
[SPARK-3237][SQL] Fix parquet filters with UDFs
Author: Michael Armbrust
Closes #2153 from marmbrus/parquetFilters and squashes the following commits:
712731a [Michael Armbrust] Use closure serializer for sending f
Repository: spark
Updated Branches:
refs/heads/branch-1.1 5cf1e4401 -> ca01de1b9
[SPARK-3237][SQL] Fix parquet filters with UDFs
Author: Michael Armbrust
Closes #2153 from marmbrus/parquetFilters and squashes the following commits:
712731a [Michael Armbrust] Use closure serializer for sendi
Repository: spark
Updated Branches:
refs/heads/master e1139dd60 -> 43dfc84f8
[SPARK-2830][MLLIB] doc update for 1.1
1. renamed mllib-basics to mllib-data-types
1. renamed mllib-stats to mllib-statistics
1. moved random data generation to the bottom of mllib-stats
1. updated toc accordingly
at
Repository: spark
Updated Branches:
refs/heads/branch-1.1 ca01de1b9 -> 74012475b
[SPARK-2830][MLLIB] doc update for 1.1
1. renamed mllib-basics to mllib-data-types
1. renamed mllib-stats to mllib-statistics
1. moved random data generation to the bottom of mllib-stats
1. updated toc accordingly
Repository: spark
Updated Branches:
refs/heads/branch-1.1 74012475b -> 7286d5707
[SPARK-3227] [mllib] Added migration guide for v1.0 to v1.1
The only updates are in DecisionTree.
CC: mengxr
Author: Joseph K. Bradley
Closes #2146 from jkbradley/mllib-migration and squashes the following com
Repository: spark
Updated Branches:
refs/heads/master 43dfc84f8 -> 171a41cb0
[SPARK-3227] [mllib] Added migration guide for v1.0 to v1.1
The only updates are in DecisionTree.
CC: mengxr
Author: Joseph K. Bradley
Closes #2146 from jkbradley/mllib-migration and squashes the following commits
Repository: spark
Updated Branches:
refs/heads/branch-1.1 7286d5707 -> 1d468df33
[SPARK-3154][STREAMING] Make FlumePollingInputDStream shutdown cleaner.
Currently lot of errors get thrown from Avro IPC layer when the dstream
or sink is shutdown. This PR cleans it up. Some refactoring is done i
Repository: spark
Updated Branches:
refs/heads/master 171a41cb0 -> 6f671d04f
[SPARK-3154][STREAMING] Make FlumePollingInputDStream shutdown cleaner.
Currently lot of errors get thrown from Avro IPC layer when the dstream
or sink is shutdown. This PR cleans it up. Some refactoring is done in th
Repository: spark
Updated Branches:
refs/heads/master 6f671d04f -> b92d823ad
http://git-wip-us.apache.org/repos/asf/spark/blob/b92d823a/yarn/common/src/main/scala/org/apache/spark/scheduler/cluster/YarnClientClusterScheduler.scala
[SPARK-2933] [yarn] Refactor and cleanup Yarn AM code.
This change modifies the Yarn module so that all the logic related
to running the ApplicationMaster is localized. Instead of, previously,
4 different classes with mostly identical code, now we have:
- A single, shared ApplicationMaster class,
Repository: spark
Updated Branches:
refs/heads/master b92d823ad -> d8298c46b
[SPARK-3170][CORE][BUG]:RDD info loss in "StorageTab" and "ExecutorTab"
compeleted stage only need to remove its own partitions that are no longer
cached. However, "StorageTab" may lost some rdds which are cached act
Repository: spark
Updated Branches:
refs/heads/branch-1.1 1d468df33 -> 8f8e2a4ee
[SPARK-3170][CORE][BUG]:RDD info loss in "StorageTab" and "ExecutorTab"
compeleted stage only need to remove its own partitions that are no longer
cached. However, "StorageTab" may lost some rdds which are cached
Repository: spark
Updated Branches:
refs/heads/branch-1.1 8f8e2a4ee -> 092121e47
[SPARK-3239] [PySpark] randomize the dirs for each process
This can avoid the IO contention during spilling, when you have multiple disks.
Author: Davies Liu
Closes #2152 from davies/randomize and squashes the
Repository: spark
Updated Branches:
refs/heads/branch-1.1 092121e47 -> 935bffe3b
[SPARK-2608][Core] Fixed command line option passing issue over Mesos via
SPARK_EXECUTOR_OPTS
This is another try after #2145 to fix
[SPARK-2608](https://issues.apache.org/jira/browse/SPARK-2608).
### Basic Ide
Repository: spark
Updated Branches:
refs/heads/branch-1.1 935bffe3b -> 0c94a5b2a
SPARK-3259 - User data should be given to the master
Author: Allan Douglas R. de Oliveira
Closes #2162 from douglaz/user_data_master and squashes the following commits:
10d15f6 [Allan Douglas R. de Oliveira] Gi
Repository: spark
Updated Branches:
refs/heads/master d8298c46b -> 5ac4093c9
SPARK-3259 - User data should be given to the master
Author: Allan Douglas R. de Oliveira
Closes #2162 from douglaz/user_data_master and squashes the following commits:
10d15f6 [Allan Douglas R. de Oliveira] Give u
Repository: spark
Updated Branches:
refs/heads/master 5ac4093c9 -> 3b5eb7083
[SPARK-3118][SQL]add "SHOW TBLPROPERTIES tblname;" and "SHOW COLUMNS (FROM|IN)
table_name [(FROM|IN) db_name]" support
JIRA issue: [SPARK-3118] https://issues.apache.org/jira/browse/SPARK-3118
eg:
> SHOW TBLPROPERTI
Repository: spark
Updated Branches:
refs/heads/branch-1.1 0c94a5b2a -> 19cda0788
[SPARK-3118][SQL]add "SHOW TBLPROPERTIES tblname;" and "SHOW COLUMNS (FROM|IN)
table_name [(FROM|IN) db_name]" support
JIRA issue: [SPARK-3118] https://issues.apache.org/jira/browse/SPARK-3118
eg:
> SHOW TBLPROP
Repository: spark
Updated Branches:
refs/heads/master 3b5eb7083 -> 4238c17dc
[SPARK-3197] [SQL] Reduce the Expression tree object creations for aggregation
function (min/max)
Aggregation function min/max in catalyst will create expression tree for each
single row, however, the expression tre
Repository: spark
Updated Branches:
refs/heads/branch-1.1 19cda0788 -> 4c7f082c6
[SPARK-3197] [SQL] Reduce the Expression tree object creations for aggregation
function (min/max)
Aggregation function min/max in catalyst will create expression tree for each
single row, however, the expression
Repository: spark
Updated Branches:
refs/heads/master 4238c17dc -> 191d7cf2a
[SPARK-3256] Added support for :cp that was broken in Scala 2.10.x for
REPL
As seen with [SI-6502](https://issues.scala-lang.org/browse/SI-6502) of Scala,
the _:cp_ command was broken in Scala 2.10.x. As the Spark
Repository: spark
Updated Branches:
refs/heads/master 191d7cf2a -> 48f42781d
[SPARK-3138][SQL] sqlContext.parquetFile should be able to take a single file
as parameter
```if (!fs.getFileStatus(path).isDir) throw Exception``` make no sense after
this commit #1370
be careful if someone is wor
Repository: spark
Updated Branches:
refs/heads/branch-1.1 4c7f082c6 -> 90f8f3eed
[SPARK-3138][SQL] sqlContext.parquetFile should be able to take a single file
as parameter
```if (!fs.getFileStatus(path).isDir) throw Exception``` make no sense after
this commit #1370
be careful if someone is
Repository: spark
Updated Branches:
refs/heads/master 48f42781d -> 4fa2fda88
[SPARK-2871] [PySpark] add RDD.lookup(key)
RDD.lookup(key)
Return the list of values in the RDD for key `key`. This operation
is done efficiently if the RDD has a known partitioner by only
sea
Repository: spark
Updated Branches:
refs/heads/branch-1.1 90f8f3eed -> 3cb4e1718
Spark-3213 Fixes issue with spark-ec2 not detecting slaves created with "Launch
More like this"
... copy the spark_cluster_tag from a spot instance requests over to the
instances.
Author: Vida Ha
Closes #2163
Repository: spark
Updated Branches:
refs/heads/master 4fa2fda88 -> 7faf755ae
Spark-3213 Fixes issue with spark-ec2 not detecting slaves created with "Launch
More like this"
... copy the spark_cluster_tag from a spot instance requests over to the
instances.
Author: Vida Ha
Closes #2163 fro
Repository: spark
Updated Branches:
refs/heads/branch-1.1 3cb4e1718 -> c1ffa3e4c
[SPARK-3243] Don't use stale spark-driver.* system properties
If we set both `spark.driver.extraClassPath` and `--driver-class-path`, then
the latter correctly overrides the former. However, the value of the syst
Repository: spark
Updated Branches:
refs/heads/master 7faf755ae -> 63a053ab1
[SPARK-3243] Don't use stale spark-driver.* system properties
If we set both `spark.driver.extraClassPath` and `--driver-class-path`, then
the latter correctly overrides the former. However, the value of the system
Repository: spark
Updated Branches:
refs/heads/master 63a053ab1 -> 28d41d627
[SPARK-3252][SQL] Add missing condition for test
According to the text message, both relations should be tested. So add the
missing condition.
Author: viirya
Closes #2159 from viirya/fix_test and squashes the foll
Repository: spark
Updated Branches:
refs/heads/branch-1.1 c1ffa3e4c -> b3d763b0b
[SPARK-3252][SQL] Add missing condition for test
According to the text message, both relations should be tested. So add the
missing condition.
Author: viirya
Closes #2159 from viirya/fix_test and squashes the
Repository: spark
Updated Branches:
refs/heads/master 28d41d627 -> cc275f4b7
[SQL] [SPARK-3236] Reading Parquet tables from Metastore mangles location
Currently we do `relation.hiveQlTable.getDataLocation.getPath`, which returns
the path-part of the URI (e.g., "s3n://my-bucket/my-path" => "/m
Repository: spark
Updated Branches:
refs/heads/branch-1.1 b3d763b0b -> 77116875f
[SQL] [SPARK-3236] Reading Parquet tables from Metastore mangles location
Currently we do `relation.hiveQlTable.getDataLocation.getPath`, which returns
the path-part of the URI (e.g., "s3n://my-bucket/my-path" =>
Repository: spark
Updated Branches:
refs/heads/master cc275f4b7 -> 65253502b
[SPARK-3065][SQL] Add locale setting to fix results do not match for
udf_unix_timestamp format " MMM dd h:mm:ss a" run with not
"America/Los_Angeles" TimeZone in HiveCompatibilitySuite
When run the udf_unix_time
Repository: spark
Updated Branches:
refs/heads/branch-1.1 77116875f -> 5ea260ebd
[SPARK-3065][SQL] Add locale setting to fix results do not match for
udf_unix_timestamp format " MMM dd h:mm:ss a" run with not
"America/Los_Angeles" TimeZone in HiveCompatibilitySuite
When run the udf_unix_
Repository: spark
Updated Branches:
refs/heads/master 65253502b -> 7d2a7a91f
[SPARK-3235][SQL] Ensure in-memory tables don't always broadcast.
Author: Michael Armbrust
Closes #2147 from marmbrus/inMemDefaultSize and squashes the following commits:
5390360 [Michael Armbrust] Merge remote-tra
Repository: spark
Updated Branches:
refs/heads/branch-1.1 5ea260ebd -> 9a62cf365
[SPARK-3235][SQL] Ensure in-memory tables don't always broadcast.
Author: Michael Armbrust
Closes #2147 from marmbrus/inMemDefaultSize and squashes the following commits:
5390360 [Michael Armbrust] Merge remote
Repository: spark
Updated Branches:
refs/heads/master 7d2a7a91f -> 8712653f1
HOTFIX: Don't build with YARN support for Mapr3
Project: http://git-wip-us.apache.org/repos/asf/spark/repo
Commit: http://git-wip-us.apache.org/repos/asf/spark/commit/8712653f
Tree: http://git-wip-us.apache.org/repos
Repository: spark
Updated Branches:
refs/heads/branch-1.1 9a62cf365 -> 0b17c7d4f
Revert "[maven-release-plugin] prepare for next development iteration"
This reverts commit 9af3fb7385d1f9f221962f1d2d725ff79bd82033.
Project: http://git-wip-us.apache.org/repos/asf/spark/repo
Commit: http://git-
Revert "[maven-release-plugin] prepare release v1.1.0-snapshot2"
This reverts commit e1535ad3c6f7400f2b7915ea91da9c60510557ba.
Project: http://git-wip-us.apache.org/repos/asf/spark/repo
Commit: http://git-wip-us.apache.org/repos/asf/spark/commit/0b17c7d4
Tree: http://git-wip-us.apache.org/repos/
Repository: spark
Updated Branches:
refs/heads/master 8712653f1 -> 64d8ecbbe
Add line continuation for script to work w/ py2.7.5
Error was -
$ SPARK_HOME=$PWD/dist ./dev/create-release/generate-changelist.py
File "./dev/create-release/generate-changelist.py", line 128
if day < SPARK_REP
Repository: spark
Updated Branches:
refs/heads/branch-1.1 0b17c7d4f -> d4cf7a068
Add line continuation for script to work w/ py2.7.5
Error was -
$ SPARK_HOME=$PWD/dist ./dev/create-release/generate-changelist.py
File "./dev/create-release/generate-changelist.py", line 128
if day < SPARK
Repository: spark
Updated Branches:
refs/heads/branch-1.1 d4cf7a068 -> 8597e9cf3
http://git-wip-us.apache.org/repos/asf/spark/blob/8597e9cf/dev/create-release/generate-changelist.py
--
diff --git a/dev/create-release/generate-c
http://git-wip-us.apache.org/repos/asf/spark/blob/8597e9cf/CHANGES.txt
--
diff --git a/CHANGES.txt b/CHANGES.txt
new file mode 100644
index 000..6efb022
--- /dev/null
+++ b/CHANGES.txt
@@ -0,0 +1,14470 @@
+Spark Change Log
+---
BUILD: Updating CHANGES.txt for Spark 1.1
Project: http://git-wip-us.apache.org/repos/asf/spark/repo
Commit: http://git-wip-us.apache.org/repos/asf/spark/commit/8597e9cf
Tree: http://git-wip-us.apache.org/repos/asf/spark/tree/8597e9cf
Diff: http://git-wip-us.apache.org/repos/asf/spark/diff/8597e9
Repository: spark
Updated Branches:
refs/heads/branch-1.1 8597e9cf3 -> 58b0be6a2
[maven-release-plugin] prepare release v1.1.0-rc1
Project: http://git-wip-us.apache.org/repos/asf/spark/repo
Commit: http://git-wip-us.apache.org/repos/asf/spark/commit/58b0be6a
Tree: http://git-wip-us.apache.org
Repository: spark
Updated Branches:
refs/heads/branch-1.1 58b0be6a2 -> 78e3c036e
[maven-release-plugin] prepare for next development iteration
Project: http://git-wip-us.apache.org/repos/asf/spark/repo
Commit: http://git-wip-us.apache.org/repos/asf/spark/commit/78e3c036
Tree: http://git-wip-u
Repository: spark
Updated Tags: refs/tags/v1.1.0-rc1 [created] 1dc825d90
-
To unsubscribe, e-mail: commits-unsubscr...@spark.apache.org
For additional commands, e-mail: commits-h...@spark.apache.org
Repository: spark
Updated Branches:
refs/heads/master 64d8ecbbe -> b86277c13
[SPARK-3271] delete unused methods in Utils
delete no used method in Utils
Author: scwf
Closes #2160 from scwf/delete-no-use-method and squashes the following commits:
d8f6b0d [scwf] delete no use method in Utils
Repository: spark
Updated Branches:
refs/heads/master b86277c13 -> f38fab97c
SPARK-3265 Allow using custom ipython executable with pyspark
Although you can make pyspark use ipython with `IPYTHON=1`, and also change the
python executable with `PYSPARK_PYTHON=...`, you can't use both at the sam
Repository: spark
Updated Branches:
refs/heads/branch-1.1 78e3c036e -> 54ccd93e6
[HOTFIX] Wait for EOF only for the PySpark shell
In `SparkSubmitDriverBootstrapper`, we wait for the parent process to send us
an `EOF` before finishing the application. This is applicable for the PySpark
shell
Repository: spark
Updated Branches:
refs/heads/master f38fab97c -> dafe34349
[HOTFIX] Wait for EOF only for the PySpark shell
In `SparkSubmitDriverBootstrapper`, we wait for the parent process to send us
an `EOF` before finishing the application. This is applicable for the PySpark
shell beca
Repository: spark
Updated Branches:
refs/heads/master dafe34349 -> 024178c57
[HOTFIX][SQL] Remove cleaning of UDFs
It is not safe to run the closure cleaner on slaves. #2153 introduced this
which broke all UDF execution on slaves. Will re-add cleaning of UDF closures
in a follow-up PR.
Au
Repository: spark
Updated Branches:
refs/heads/branch-1.1 54ccd93e6 -> 233c283e3
[HOTFIX][SQL] Remove cleaning of UDFs
It is not safe to run the closure cleaner on slaves. #2153 introduced this
which broke all UDF execution on slaves. Will re-add cleaning of UDF closures
in a follow-up PR.
Repository: spark
Updated Branches:
refs/heads/branch-1.1 233c283e3 -> ad0fab23d
HOTFIX: Don't build with YARN support for Mapr3
Project: http://git-wip-us.apache.org/repos/asf/spark/repo
Commit: http://git-wip-us.apache.org/repos/asf/spark/commit/ad0fab23
Tree: http://git-wip-us.apache.org/r
Repository: spark
Updated Tags: refs/tags/v1.1.0-rc1 [deleted] 1dc825d90
-
To unsubscribe, e-mail: commits-unsubscr...@spark.apache.org
For additional commands, e-mail: commits-h...@spark.apache.org
Repository: spark
Updated Branches:
refs/heads/branch-1.1 79e86ef3e -> a118ea5c5
[maven-release-plugin] prepare for next development iteration
Project: http://git-wip-us.apache.org/repos/asf/spark/repo
Commit: http://git-wip-us.apache.org/repos/asf/spark/commit/a118ea5c
Tree: http://git-wip-u
Repository: spark
Updated Tags: refs/tags/v1.1.0-rc1 [created] d6d19b581
-
To unsubscribe, e-mail: commits-unsubscr...@spark.apache.org
For additional commands, e-mail: commits-h...@spark.apache.org
Repository: spark
Updated Branches:
refs/heads/branch-1.1 ad0fab23d -> 79e86ef3e
[maven-release-plugin] prepare release v1.1.0-rc1
Project: http://git-wip-us.apache.org/repos/asf/spark/repo
Commit: http://git-wip-us.apache.org/repos/asf/spark/commit/79e86ef3
Tree: http://git-wip-us.apache.org
60 matches
Mail list logo