git commit: [SPARK-1275] Made dev/run-tests executable.

2014-03-19 Thread tdas
Repository: spark
Updated Branches:
  refs/heads/branch-0.9 72875b29b - a4eef655c


[SPARK-1275] Made dev/run-tests executable.

This was causing Jenkins tests to fail for PRs against branch 0.9.

Author: Tathagata Das tathagata.das1...@gmail.com

Closes #178 from tdas/branch-0.9-fix and squashes the following commits:

a633bce [Tathagata Das] Merge remote-tracking branch 'apache-github/branch-0.9' 
into branch-0.9-fix
9b043cc [Tathagata Das] Made dev/run-tests executable.


Project: http://git-wip-us.apache.org/repos/asf/spark/repo
Commit: http://git-wip-us.apache.org/repos/asf/spark/commit/a4eef655
Tree: http://git-wip-us.apache.org/repos/asf/spark/tree/a4eef655
Diff: http://git-wip-us.apache.org/repos/asf/spark/diff/a4eef655

Branch: refs/heads/branch-0.9
Commit: a4eef655c7263f73db6f2c837288982140f17a23
Parents: 72875b2
Author: Tathagata Das tathagata.das1...@gmail.com
Authored: Wed Mar 19 16:10:45 2014 -0700
Committer: Tathagata Das tathagata.das1...@gmail.com
Committed: Wed Mar 19 16:10:45 2014 -0700

--
 dev/run-tests | 0
 1 file changed, 0 insertions(+), 0 deletions(-)
--


http://git-wip-us.apache.org/repos/asf/spark/blob/a4eef655/dev/run-tests
--
diff --git a/dev/run-tests b/dev/run-tests
old mode 100644
new mode 100755



git commit: Bumped versions to Spark 0.9.1

2014-03-20 Thread tdas
Repository: spark
Updated Branches:
  refs/heads/branch-0.9 c6630d363 - 8a882eff9


Bumped versions to Spark 0.9.1

Self explanatory!

Author: Tathagata Das tathagata.das1...@gmail.com

Closes #190 from tdas/branch-0.9-version-bump and squashes the following 
commits:

34576ee [Tathagata Das] Bumped versions to Spark 0.9.1


Project: http://git-wip-us.apache.org/repos/asf/spark/repo
Commit: http://git-wip-us.apache.org/repos/asf/spark/commit/8a882eff
Tree: http://git-wip-us.apache.org/repos/asf/spark/tree/8a882eff
Diff: http://git-wip-us.apache.org/repos/asf/spark/diff/8a882eff

Branch: refs/heads/branch-0.9
Commit: 8a882eff9d4a7377ba2f470a59a3ab4722c66d9d
Parents: c6630d3
Author: Tathagata Das tathagata.das1...@gmail.com
Authored: Thu Mar 20 16:55:35 2014 -0700
Committer: Tathagata Das tathagata.das1...@gmail.com
Committed: Thu Mar 20 16:55:35 2014 -0700

--
 docs/_config.yml   | 4 ++--
 ec2/spark_ec2.py   | 4 ++--
 pom.xml| 2 +-
 project/SparkBuild.scala   | 2 +-
 python/pyspark/shell.py| 2 +-
 repl/src/main/scala/org/apache/spark/repl/SparkILoopInit.scala | 2 +-
 6 files changed, 8 insertions(+), 8 deletions(-)
--


http://git-wip-us.apache.org/repos/asf/spark/blob/8a882eff/docs/_config.yml
--
diff --git a/docs/_config.yml b/docs/_config.yml
index a7f46ff..db6cf74 100644
--- a/docs/_config.yml
+++ b/docs/_config.yml
@@ -3,8 +3,8 @@ markdown: kramdown
 
 # These allow the documentation to be updated with nerw releases
 # of Spark, Scala, and Mesos.
-SPARK_VERSION: 0.9.0-incubating
-SPARK_VERSION_SHORT: 0.9.0
+SPARK_VERSION: 0.9.1
+SPARK_VERSION_SHORT: 0.9.1
 SCALA_BINARY_VERSION: 2.10
 SCALA_VERSION: 2.10.3
 MESOS_VERSION: 0.13.0

http://git-wip-us.apache.org/repos/asf/spark/blob/8a882eff/ec2/spark_ec2.py
--
diff --git a/ec2/spark_ec2.py b/ec2/spark_ec2.py
index e88f80a..ca944ce 100755
--- a/ec2/spark_ec2.py
+++ b/ec2/spark_ec2.py
@@ -70,7 +70,7 @@ def parse_args():
slaves across multiple (an additional $0.01/Gb for bandwidth +
between zones applies))
   parser.add_option(-a, --ami, help=Amazon Machine Image ID to use)
-  parser.add_option(-v, --spark-version, default=0.9.0,
+  parser.add_option(-v, --spark-version, default=0.9.1,
   help=Version of Spark to use: 'X.Y.Z' or a specific git hash)
   parser.add_option(--spark-git-repo,
   default=https://github.com/apache/incubator-spark;,
@@ -157,7 +157,7 @@ def is_active(instance):
 
 # Return correct versions of Spark and Shark, given the supplied Spark version
 def get_spark_shark_version(opts):
-  spark_shark_map = {0.7.3: 0.7.1, 0.8.0: 0.8.0, 0.8.1: 0.8.1, 
0.9.0: 0.9.0}
+  spark_shark_map = {0.7.3: 0.7.1, 0.8.0: 0.8.0, 0.8.1: 0.8.1, 
0.9.0: 0.9.0, 0.9.1: 0.9.1}
   version = opts.spark_version.replace(v, )
   if version not in spark_shark_map:
 print  stderr, Don't know about Spark version: %s % version

http://git-wip-us.apache.org/repos/asf/spark/blob/8a882eff/pom.xml
--
diff --git a/pom.xml b/pom.xml
index e53c930..f7ee914 100644
--- a/pom.xml
+++ b/pom.xml
@@ -25,7 +25,7 @@
   /parent
   groupIdorg.apache.spark/groupId
   artifactIdspark-parent/artifactId
-  version0.9.1-incubating-SNAPSHOT/version
+  version0.9.1-SNAPSHOT/version
   packagingpom/packaging
   nameSpark Project Parent POM/name
   urlhttp://spark.incubator.apache.org//url

http://git-wip-us.apache.org/repos/asf/spark/blob/8a882eff/project/SparkBuild.scala
--
diff --git a/project/SparkBuild.scala b/project/SparkBuild.scala
index f9eeeb0..c5d6327 100644
--- a/project/SparkBuild.scala
+++ b/project/SparkBuild.scala
@@ -134,7 +134,7 @@ object SparkBuild extends Build {
 
   def sharedSettings = Defaults.defaultSettings ++ Seq(
 organization   := org.apache.spark,
-version:= 0.9.0-incubating,
+version:= 0.9.1,
 scalaVersion   := 2.10.3,
 scalacOptions := Seq(-Xmax-classfile-name, 120, -unchecked, 
-deprecation,
   -target: + SCALAC_JVM_VERSION),

http://git-wip-us.apache.org/repos/asf/spark/blob/8a882eff/python/pyspark/shell.py
--
diff --git a/python/pyspark/shell.py b/python/pyspark/shell.py
index 9203342..0425192 100644
--- a/python/pyspark/shell.py
+++ b/python/pyspark/shell.py
@@ -35,7 +35,7 @@ print Welcome

[1/3] git commit: Revert [maven-release-plugin] prepare for next development iteration

2014-03-26 Thread tdas
Repository: spark
Updated Branches:
  refs/heads/branch-0.9 49016046d - 348f54b9c


Revert [maven-release-plugin] prepare for next development iteration

This reverts commit 1f785d46e4e3df716dc836e38933dc0a30028496.


Project: http://git-wip-us.apache.org/repos/asf/spark/repo
Commit: http://git-wip-us.apache.org/repos/asf/spark/commit/d16e863d
Tree: http://git-wip-us.apache.org/repos/asf/spark/tree/d16e863d
Diff: http://git-wip-us.apache.org/repos/asf/spark/diff/d16e863d

Branch: refs/heads/branch-0.9
Commit: d16e863de400d47bbf2b6dc6ee8d9d67db7cd02a
Parents: 4901604
Author: Tathagata Das tathagata.das1...@gmail.com
Authored: Wed Mar 26 21:51:40 2014 -0700
Committer: Tathagata Das tathagata.das1...@gmail.com
Committed: Wed Mar 26 21:51:40 2014 -0700

--
 assembly/pom.xml  | 2 +-
 bagel/pom.xml | 2 +-
 core/pom.xml  | 2 +-
 examples/pom.xml  | 2 +-
 external/flume/pom.xml| 2 +-
 external/kafka/pom.xml| 2 +-
 external/mqtt/pom.xml | 2 +-
 external/twitter/pom.xml  | 2 +-
 external/zeromq/pom.xml   | 2 +-
 extras/spark-ganglia-lgpl/pom.xml | 2 +-
 graphx/pom.xml| 2 +-
 mllib/pom.xml | 2 +-
 pom.xml   | 4 ++--
 repl/pom.xml  | 2 +-
 streaming/pom.xml | 2 +-
 tools/pom.xml | 2 +-
 yarn/pom.xml  | 2 +-
 yarn/stable/pom.xml   | 2 +-
 18 files changed, 19 insertions(+), 19 deletions(-)
--


http://git-wip-us.apache.org/repos/asf/spark/blob/d16e863d/assembly/pom.xml
--
diff --git a/assembly/pom.xml b/assembly/pom.xml
index 0edbdd6..eb884b3 100644
--- a/assembly/pom.xml
+++ b/assembly/pom.xml
@@ -21,7 +21,7 @@
   parent
 groupIdorg.apache.spark/groupId
 artifactIdspark-parent/artifactId
-version0.9.2-SNAPSHOT/version
+version0.9.1/version
 relativePath../pom.xml/relativePath
   /parent
 

http://git-wip-us.apache.org/repos/asf/spark/blob/d16e863d/bagel/pom.xml
--
diff --git a/bagel/pom.xml b/bagel/pom.xml
index b4b5588..4ca574d 100644
--- a/bagel/pom.xml
+++ b/bagel/pom.xml
@@ -21,7 +21,7 @@
   parent
 groupIdorg.apache.spark/groupId
 artifactIdspark-parent/artifactId
-version0.9.2-SNAPSHOT/version
+version0.9.1/version
 relativePath../pom.xml/relativePath
   /parent
 

http://git-wip-us.apache.org/repos/asf/spark/blob/d16e863d/core/pom.xml
--
diff --git a/core/pom.xml b/core/pom.xml
index 0fdf380..8111400 100644
--- a/core/pom.xml
+++ b/core/pom.xml
@@ -21,7 +21,7 @@
 parent
 groupIdorg.apache.spark/groupId
 artifactIdspark-parent/artifactId
-version0.9.2-SNAPSHOT/version
+version0.9.1/version
 relativePath../pom.xml/relativePath
 /parent
 

http://git-wip-us.apache.org/repos/asf/spark/blob/d16e863d/examples/pom.xml
--
diff --git a/examples/pom.xml b/examples/pom.xml
index 831a2df..9ae69e9 100644
--- a/examples/pom.xml
+++ b/examples/pom.xml
@@ -21,7 +21,7 @@
   parent
 groupIdorg.apache.spark/groupId
 artifactIdspark-parent/artifactId
-version0.9.2-SNAPSHOT/version
+version0.9.1/version
 relativePath../pom.xml/relativePath
   /parent
 

http://git-wip-us.apache.org/repos/asf/spark/blob/d16e863d/external/flume/pom.xml
--
diff --git a/external/flume/pom.xml b/external/flume/pom.xml
index fcd19e7..655fed1 100644
--- a/external/flume/pom.xml
+++ b/external/flume/pom.xml
@@ -21,7 +21,7 @@
   parent
 groupIdorg.apache.spark/groupId
 artifactIdspark-parent/artifactId
-version0.9.2-SNAPSHOT/version
+version0.9.1/version
 relativePath../../pom.xml/relativePath
   /parent
 

http://git-wip-us.apache.org/repos/asf/spark/blob/d16e863d/external/kafka/pom.xml
--
diff --git a/external/kafka/pom.xml b/external/kafka/pom.xml
index e495877..0814d33 100644
--- a/external/kafka/pom.xml
+++ b/external/kafka/pom.xml
@@ -21,7 +21,7 @@
   parent
 groupIdorg.apache.spark/groupId
 artifactIdspark-parent/artifactId
-version0.9.2-SNAPSHOT/version
+version0.9.1/version
 relativePath../../pom.xml/relativePath
   /parent
 

http://git-wip-us.apache.org/repos/asf/spark/blob/d16e863d/external/mqtt/pom.xml
--
diff --git a/external/mqtt/pom.xml b/external/mqtt/pom.xml
index 33e3077..7a67e14 100644
--- a/external/mqtt/pom.xml
+++ 

[2/3] git commit: Revert [maven-release-plugin] prepare release v0.9.1-rc2

2014-03-26 Thread tdas
Revert [maven-release-plugin] prepare release v0.9.1-rc2

This reverts commit 1197280acf1322165301259dd825f44e22a323bc.


Project: http://git-wip-us.apache.org/repos/asf/spark/repo
Commit: http://git-wip-us.apache.org/repos/asf/spark/commit/ea5da04f
Tree: http://git-wip-us.apache.org/repos/asf/spark/tree/ea5da04f
Diff: http://git-wip-us.apache.org/repos/asf/spark/diff/ea5da04f

Branch: refs/heads/branch-0.9
Commit: ea5da04f77b1de55a1d3ea03afef5dbc8c298894
Parents: d16e863
Author: Tathagata Das tathagata.das1...@gmail.com
Authored: Wed Mar 26 21:53:07 2014 -0700
Committer: Tathagata Das tathagata.das1...@gmail.com
Committed: Wed Mar 26 21:53:07 2014 -0700

--
 assembly/pom.xml  | 2 +-
 bagel/pom.xml | 2 +-
 core/pom.xml  | 2 +-
 examples/pom.xml  | 2 +-
 external/flume/pom.xml| 2 +-
 external/kafka/pom.xml| 2 +-
 external/mqtt/pom.xml | 2 +-
 external/twitter/pom.xml  | 2 +-
 external/zeromq/pom.xml   | 2 +-
 extras/spark-ganglia-lgpl/pom.xml | 2 +-
 graphx/pom.xml| 2 +-
 mllib/pom.xml | 2 +-
 pom.xml   | 4 ++--
 repl/pom.xml  | 2 +-
 streaming/pom.xml | 2 +-
 tools/pom.xml | 2 +-
 yarn/pom.xml  | 2 +-
 yarn/stable/pom.xml   | 2 +-
 18 files changed, 19 insertions(+), 19 deletions(-)
--


http://git-wip-us.apache.org/repos/asf/spark/blob/ea5da04f/assembly/pom.xml
--
diff --git a/assembly/pom.xml b/assembly/pom.xml
index eb884b3..d22ffc8 100644
--- a/assembly/pom.xml
+++ b/assembly/pom.xml
@@ -21,7 +21,7 @@
   parent
 groupIdorg.apache.spark/groupId
 artifactIdspark-parent/artifactId
-version0.9.1/version
+version0.9.1-SNAPSHOT/version
 relativePath../pom.xml/relativePath
   /parent
 

http://git-wip-us.apache.org/repos/asf/spark/blob/ea5da04f/bagel/pom.xml
--
diff --git a/bagel/pom.xml b/bagel/pom.xml
index 4ca574d..4a86e1b6 100644
--- a/bagel/pom.xml
+++ b/bagel/pom.xml
@@ -21,7 +21,7 @@
   parent
 groupIdorg.apache.spark/groupId
 artifactIdspark-parent/artifactId
-version0.9.1/version
+version0.9.1-SNAPSHOT/version
 relativePath../pom.xml/relativePath
   /parent
 

http://git-wip-us.apache.org/repos/asf/spark/blob/ea5da04f/core/pom.xml
--
diff --git a/core/pom.xml b/core/pom.xml
index 8111400..aea662a 100644
--- a/core/pom.xml
+++ b/core/pom.xml
@@ -21,7 +21,7 @@
 parent
 groupIdorg.apache.spark/groupId
 artifactIdspark-parent/artifactId
-version0.9.1/version
+version0.9.1-SNAPSHOT/version
 relativePath../pom.xml/relativePath
 /parent
 

http://git-wip-us.apache.org/repos/asf/spark/blob/ea5da04f/examples/pom.xml
--
diff --git a/examples/pom.xml b/examples/pom.xml
index 9ae69e9..568f55e 100644
--- a/examples/pom.xml
+++ b/examples/pom.xml
@@ -21,7 +21,7 @@
   parent
 groupIdorg.apache.spark/groupId
 artifactIdspark-parent/artifactId
-version0.9.1/version
+version0.9.1-SNAPSHOT/version
 relativePath../pom.xml/relativePath
   /parent
 

http://git-wip-us.apache.org/repos/asf/spark/blob/ea5da04f/external/flume/pom.xml
--
diff --git a/external/flume/pom.xml b/external/flume/pom.xml
index 655fed1..8896477 100644
--- a/external/flume/pom.xml
+++ b/external/flume/pom.xml
@@ -21,7 +21,7 @@
   parent
 groupIdorg.apache.spark/groupId
 artifactIdspark-parent/artifactId
-version0.9.1/version
+version0.9.1-SNAPSHOT/version
 relativePath../../pom.xml/relativePath
   /parent
 

http://git-wip-us.apache.org/repos/asf/spark/blob/ea5da04f/external/kafka/pom.xml
--
diff --git a/external/kafka/pom.xml b/external/kafka/pom.xml
index 0814d33..8d5ebb5 100644
--- a/external/kafka/pom.xml
+++ b/external/kafka/pom.xml
@@ -21,7 +21,7 @@
   parent
 groupIdorg.apache.spark/groupId
 artifactIdspark-parent/artifactId
-version0.9.1/version
+version0.9.1-SNAPSHOT/version
 relativePath../../pom.xml/relativePath
   /parent
 

http://git-wip-us.apache.org/repos/asf/spark/blob/ea5da04f/external/mqtt/pom.xml
--
diff --git a/external/mqtt/pom.xml b/external/mqtt/pom.xml
index 7a67e14..61092e5 100644
--- a/external/mqtt/pom.xml
+++ b/external/mqtt/pom.xml
@@ -21,7 +21,7 @@
   parent
 groupIdorg.apache.spark/groupId
 

git commit: [maven-release-plugin] prepare release v0.9.1-rc3

2014-03-26 Thread tdas
Repository: spark
Updated Branches:
  refs/heads/branch-0.9 348f54b9c - 4c43182b6


[maven-release-plugin] prepare release v0.9.1-rc3


Project: http://git-wip-us.apache.org/repos/asf/spark/repo
Commit: http://git-wip-us.apache.org/repos/asf/spark/commit/4c43182b
Tree: http://git-wip-us.apache.org/repos/asf/spark/tree/4c43182b
Diff: http://git-wip-us.apache.org/repos/asf/spark/diff/4c43182b

Branch: refs/heads/branch-0.9
Commit: 4c43182b6d1b0b7717423f386c0214fe93073208
Parents: 348f54b
Author: Tathagata Das tathagata.das1...@gmail.com
Authored: Thu Mar 27 05:14:46 2014 +
Committer: Tathagata Das tathagata.das1...@gmail.com
Committed: Thu Mar 27 05:14:46 2014 +

--
 assembly/pom.xml  | 2 +-
 bagel/pom.xml | 2 +-
 core/pom.xml  | 2 +-
 examples/pom.xml  | 2 +-
 external/flume/pom.xml| 2 +-
 external/kafka/pom.xml| 2 +-
 external/mqtt/pom.xml | 2 +-
 external/twitter/pom.xml  | 2 +-
 external/zeromq/pom.xml   | 2 +-
 extras/spark-ganglia-lgpl/pom.xml | 2 +-
 graphx/pom.xml| 2 +-
 mllib/pom.xml | 2 +-
 pom.xml   | 4 ++--
 repl/pom.xml  | 2 +-
 streaming/pom.xml | 2 +-
 tools/pom.xml | 2 +-
 yarn/pom.xml  | 2 +-
 yarn/stable/pom.xml   | 2 +-
 18 files changed, 19 insertions(+), 19 deletions(-)
--


http://git-wip-us.apache.org/repos/asf/spark/blob/4c43182b/assembly/pom.xml
--
diff --git a/assembly/pom.xml b/assembly/pom.xml
index d22ffc8..eb884b3 100644
--- a/assembly/pom.xml
+++ b/assembly/pom.xml
@@ -21,7 +21,7 @@
   parent
 groupIdorg.apache.spark/groupId
 artifactIdspark-parent/artifactId
-version0.9.1-SNAPSHOT/version
+version0.9.1/version
 relativePath../pom.xml/relativePath
   /parent
 

http://git-wip-us.apache.org/repos/asf/spark/blob/4c43182b/bagel/pom.xml
--
diff --git a/bagel/pom.xml b/bagel/pom.xml
index 4a86e1b6..4ca574d 100644
--- a/bagel/pom.xml
+++ b/bagel/pom.xml
@@ -21,7 +21,7 @@
   parent
 groupIdorg.apache.spark/groupId
 artifactIdspark-parent/artifactId
-version0.9.1-SNAPSHOT/version
+version0.9.1/version
 relativePath../pom.xml/relativePath
   /parent
 

http://git-wip-us.apache.org/repos/asf/spark/blob/4c43182b/core/pom.xml
--
diff --git a/core/pom.xml b/core/pom.xml
index aea662a..8111400 100644
--- a/core/pom.xml
+++ b/core/pom.xml
@@ -21,7 +21,7 @@
 parent
 groupIdorg.apache.spark/groupId
 artifactIdspark-parent/artifactId
-version0.9.1-SNAPSHOT/version
+version0.9.1/version
 relativePath../pom.xml/relativePath
 /parent
 

http://git-wip-us.apache.org/repos/asf/spark/blob/4c43182b/examples/pom.xml
--
diff --git a/examples/pom.xml b/examples/pom.xml
index 568f55e..9ae69e9 100644
--- a/examples/pom.xml
+++ b/examples/pom.xml
@@ -21,7 +21,7 @@
   parent
 groupIdorg.apache.spark/groupId
 artifactIdspark-parent/artifactId
-version0.9.1-SNAPSHOT/version
+version0.9.1/version
 relativePath../pom.xml/relativePath
   /parent
 

http://git-wip-us.apache.org/repos/asf/spark/blob/4c43182b/external/flume/pom.xml
--
diff --git a/external/flume/pom.xml b/external/flume/pom.xml
index 8896477..655fed1 100644
--- a/external/flume/pom.xml
+++ b/external/flume/pom.xml
@@ -21,7 +21,7 @@
   parent
 groupIdorg.apache.spark/groupId
 artifactIdspark-parent/artifactId
-version0.9.1-SNAPSHOT/version
+version0.9.1/version
 relativePath../../pom.xml/relativePath
   /parent
 

http://git-wip-us.apache.org/repos/asf/spark/blob/4c43182b/external/kafka/pom.xml
--
diff --git a/external/kafka/pom.xml b/external/kafka/pom.xml
index 8d5ebb5..0814d33 100644
--- a/external/kafka/pom.xml
+++ b/external/kafka/pom.xml
@@ -21,7 +21,7 @@
   parent
 groupIdorg.apache.spark/groupId
 artifactIdspark-parent/artifactId
-version0.9.1-SNAPSHOT/version
+version0.9.1/version
 relativePath../../pom.xml/relativePath
   /parent
 

http://git-wip-us.apache.org/repos/asf/spark/blob/4c43182b/external/mqtt/pom.xml
--
diff --git a/external/mqtt/pom.xml b/external/mqtt/pom.xml
index 61092e5..7a67e14 100644
--- a/external/mqtt/pom.xml
+++ b/external/mqtt/pom.xml
@@ -21,7 +21,7 @@
   parent
 

svn commit: r1585872 - in /spark/site/docs/0.9.1: ./ api/ api/bagel/ api/bagel/index/ api/bagel/lib/ api/bagel/org/ api/bagel/org/apache/ api/bagel/org/apache/spark/ api/bagel/org/apache/spark/bagel/

2014-04-08 Thread tdas
Author: tdas
Date: Wed Apr  9 01:34:23 2014
New Revision: 1585872

URL: http://svn.apache.org/r1585872
Log:
Added 0.9.1 docs


[This commit notification would consist of 309 parts, 
which exceeds the limit of 50 ones, so it was shortened to the summary.]


Git Push Summary

2014-04-08 Thread tdas
Repository: spark
Updated Tags:  refs/tags/v0.9.1 [created] 4c43182b6


svn commit: r5003 - /dev/spark/spark-0.9.1/

2014-04-09 Thread tdas
Author: tdas
Date: Wed Apr  9 19:48:15 2014
New Revision: 5003

Log:
Moved Spark 0.9.1 from dev tree to release tree, within incubator/spark as a 
temporary place
until www.apache.org/dist/spark is created.

Removed:
dev/spark/spark-0.9.1/



svn commit: r1586127 [2/2] - in /spark: ./ news/_posts/ releases/_posts/ site/ site/docs/ site/mllib/ site/news/ site/releases/ site/screencasts/ site/streaming/

2014-04-09 Thread tdas
Modified: spark/site/releases/spark-release-0-5-0.html
URL: 
http://svn.apache.org/viewvc/spark/site/releases/spark-release-0-5-0.html?rev=1586127r1=1586126r2=1586127view=diff
==
--- spark/site/releases/spark-release-0-5-0.html (original)
+++ spark/site/releases/spark-release-0-5-0.html Wed Apr  9 20:17:50 2014
@@ -124,6 +124,9 @@
   h5Latest News/h5
   ul class=list-unstyled
 
+  lia href=/news/spark-0-9-1-released.htmlSpark 0.9.1 
released/a
+  span class=small(Apr 09, 2014)/span/li
+
   lia 
href=/news/submit-talks-to-spark-summit-2014.htmlSubmissions and 
registration open for Spark Summit 2014/a
   span class=small(Mar 20, 2014)/span/li
 
@@ -133,9 +136,6 @@
   lia href=/news/spark-0-9-0-released.htmlSpark 0.9.0 
released/a
   span class=small(Feb 02, 2014)/span/li
 
-  lia href=/news/spark-0-8-1-released.htmlSpark 0.8.1 
released/a
-  span class=small(Dec 19, 2013)/span/li
-
   /ul
   p class=small style=text-align: right;a 
href=/news/index.htmlArchive/a/p
 /div

Modified: spark/site/releases/spark-release-0-5-1.html
URL: 
http://svn.apache.org/viewvc/spark/site/releases/spark-release-0-5-1.html?rev=1586127r1=1586126r2=1586127view=diff
==
--- spark/site/releases/spark-release-0-5-1.html (original)
+++ spark/site/releases/spark-release-0-5-1.html Wed Apr  9 20:17:50 2014
@@ -124,6 +124,9 @@
   h5Latest News/h5
   ul class=list-unstyled
 
+  lia href=/news/spark-0-9-1-released.htmlSpark 0.9.1 
released/a
+  span class=small(Apr 09, 2014)/span/li
+
   lia 
href=/news/submit-talks-to-spark-summit-2014.htmlSubmissions and 
registration open for Spark Summit 2014/a
   span class=small(Mar 20, 2014)/span/li
 
@@ -133,9 +136,6 @@
   lia href=/news/spark-0-9-0-released.htmlSpark 0.9.0 
released/a
   span class=small(Feb 02, 2014)/span/li
 
-  lia href=/news/spark-0-8-1-released.htmlSpark 0.8.1 
released/a
-  span class=small(Dec 19, 2013)/span/li
-
   /ul
   p class=small style=text-align: right;a 
href=/news/index.htmlArchive/a/p
 /div

Modified: spark/site/releases/spark-release-0-5-2.html
URL: 
http://svn.apache.org/viewvc/spark/site/releases/spark-release-0-5-2.html?rev=1586127r1=1586126r2=1586127view=diff
==
--- spark/site/releases/spark-release-0-5-2.html (original)
+++ spark/site/releases/spark-release-0-5-2.html Wed Apr  9 20:17:50 2014
@@ -124,6 +124,9 @@
   h5Latest News/h5
   ul class=list-unstyled
 
+  lia href=/news/spark-0-9-1-released.htmlSpark 0.9.1 
released/a
+  span class=small(Apr 09, 2014)/span/li
+
   lia 
href=/news/submit-talks-to-spark-summit-2014.htmlSubmissions and 
registration open for Spark Summit 2014/a
   span class=small(Mar 20, 2014)/span/li
 
@@ -133,9 +136,6 @@
   lia href=/news/spark-0-9-0-released.htmlSpark 0.9.0 
released/a
   span class=small(Feb 02, 2014)/span/li
 
-  lia href=/news/spark-0-8-1-released.htmlSpark 0.8.1 
released/a
-  span class=small(Dec 19, 2013)/span/li
-
   /ul
   p class=small style=text-align: right;a 
href=/news/index.htmlArchive/a/p
 /div

Modified: spark/site/releases/spark-release-0-6-0.html
URL: 
http://svn.apache.org/viewvc/spark/site/releases/spark-release-0-6-0.html?rev=1586127r1=1586126r2=1586127view=diff
==
--- spark/site/releases/spark-release-0-6-0.html (original)
+++ spark/site/releases/spark-release-0-6-0.html Wed Apr  9 20:17:50 2014
@@ -124,6 +124,9 @@
   h5Latest News/h5
   ul class=list-unstyled
 
+  lia href=/news/spark-0-9-1-released.htmlSpark 0.9.1 
released/a
+  span class=small(Apr 09, 2014)/span/li
+
   lia 
href=/news/submit-talks-to-spark-summit-2014.htmlSubmissions and 
registration open for Spark Summit 2014/a
   span class=small(Mar 20, 2014)/span/li
 
@@ -133,9 +136,6 @@
   lia href=/news/spark-0-9-0-released.htmlSpark 0.9.0 
released/a
   span class=small(Feb 02, 2014)/span/li
 
-  lia href=/news/spark-0-8-1-released.htmlSpark 0.8.1 
released/a
-  span class=small(Dec 19, 2013)/span/li
-
   /ul
   p class=small style=text-align: right;a 
href=/news/index.htmlArchive/a/p
 /div

Modified: spark/site/releases/spark-release-0-6-1.html
URL: 
http://svn.apache.org/viewvc/spark/site/releases/spark-release-0-6-1.html?rev=1586127r1=1586126r2=1586127view=diff
==
--- 

svn commit: r1586167 [2/2] - in /spark/site/docs/0.9.1: ./ api/pyspark/

2014-04-09 Thread tdas
Modified: spark/site/docs/0.9.1/tuning.html
URL: 
http://svn.apache.org/viewvc/spark/site/docs/0.9.1/tuning.html?rev=1586167r1=1586166r2=1586167view=diff
==
--- spark/site/docs/0.9.1/tuning.html (original)
+++ spark/site/docs/0.9.1/tuning.html Wed Apr  9 22:56:47 2014
@@ -25,6 +25,19 @@
 link rel=stylesheet href=css/pygments-default.css
 
 
+!-- Google analytics script --
+script type=text/javascript
+  var _gaq = _gaq || [];
+  _gaq.push(['_setAccount', 'UA-32518208-1']);
+  _gaq.push(['_trackPageview']);
+
+  (function() {
+var ga = document.createElement('script'); ga.type = 
'text/javascript'; ga.async = true;
+ga.src = ('https:' == document.location.protocol ? 'https://ssl' : 
'http://www') + '.google-analytics.com/ga.js';
+var s = document.getElementsByTagName('script')[0]; 
s.parentNode.insertBefore(ga, s);
+  })();
+/script
+
 
 /head
 body




svn commit: r1586169 - in /spark: releases/_posts/2014-04-09-spark-release-0-9-1.md site/docs/0.9.1/0.9.1-docs.tgz site/releases/spark-release-0-9-1.html

2014-04-09 Thread tdas
Author: tdas
Date: Wed Apr  9 23:19:13 2014
New Revision: 1586169

URL: http://svn.apache.org/r1586169
Log:
Updated 0.9.1 release notes

Removed:
spark/site/docs/0.9.1/0.9.1-docs.tgz
Modified:
spark/releases/_posts/2014-04-09-spark-release-0-9-1.md
spark/site/releases/spark-release-0-9-1.html

Modified: spark/releases/_posts/2014-04-09-spark-release-0-9-1.md
URL: 
http://svn.apache.org/viewvc/spark/releases/_posts/2014-04-09-spark-release-0-9-1.md?rev=1586169r1=1586168r2=1586169view=diff
==
--- spark/releases/_posts/2014-04-09-spark-release-0-9-1.md (original)
+++ spark/releases/_posts/2014-04-09-spark-release-0-9-1.md Wed Apr  9 23:19:13 
2014
@@ -59,7 +59,7 @@ Several bug fixes were made to YARN depl
 * Fixed bug in Python de-pickling 
[[SPARK-1135](https://issues.apache.org/jira/browse/SPARK-1135)]
 * Fixed bug in serialization of strings longer than 64K 
[[SPARK-1043](https://issues.apache.org/jira/browse/SPARK-1043)] 
 * Fixed bug that made jobs hang when base file is not available 
[[SPARK-1025](https://issues.apache.org/jira/browse/SPARK-1025)] 
-* Added Missing RDD operations to PySpark - top, zip, foldByKey and setName 
+* Added Missing RDD operations to PySpark - top, zip, foldByKey, repartition, 
coallesce, getStorageLevel, setName and toDebugString
 
 ### Improvements to documentation
 * Streaming: Added documentation on running streaming application from 
spark-shell

Modified: spark/site/releases/spark-release-0-9-1.html
URL: 
http://svn.apache.org/viewvc/spark/site/releases/spark-release-0-9-1.html?rev=1586169r1=1586168r2=1586169view=diff
==
--- spark/site/releases/spark-release-0-9-1.html (original)
+++ spark/site/releases/spark-release-0-9-1.html Wed Apr  9 23:19:13 2014
@@ -217,7 +217,7 @@
   liFixed bug in Python de-pickling [a 
href=https://issues.apache.org/jira/browse/SPARK-1135;SPARK-1135/a]/li
   liFixed bug in serialization of strings longer than 64K [a 
href=https://issues.apache.org/jira/browse/SPARK-1043;SPARK-1043/a] /li
   liFixed bug that made jobs hang when base file is not available [a 
href=https://issues.apache.org/jira/browse/SPARK-1025;SPARK-1025/a] /li
-  liAdded Missing RDD operations to PySpark - top, zip, foldByKey and 
setName /li
+  liAdded Missing RDD operations to PySpark - top, zip, foldByKey, 
repartition, coallesce, getStorageLevel, setName and toDebugString/li
 /ul
 
 h3 id=improvements-to-documentationImprovements to documentation/h3




svn commit: r1586197 - in /spark: releases/_posts/2014-04-09-spark-release-0-9-1.md site/releases/spark-release-0-9-1.html

2014-04-09 Thread tdas
Author: tdas
Date: Thu Apr 10 02:17:44 2014
New Revision: 1586197

URL: http://svn.apache.org/r1586197
Log:
Fixed typo

Modified:
spark/releases/_posts/2014-04-09-spark-release-0-9-1.md
spark/site/releases/spark-release-0-9-1.html

Modified: spark/releases/_posts/2014-04-09-spark-release-0-9-1.md
URL: 
http://svn.apache.org/viewvc/spark/releases/_posts/2014-04-09-spark-release-0-9-1.md?rev=1586197r1=1586196r2=1586197view=diff
==
--- spark/releases/_posts/2014-04-09-spark-release-0-9-1.md (original)
+++ spark/releases/_posts/2014-04-09-spark-release-0-9-1.md Thu Apr 10 02:17:44 
2014
@@ -59,7 +59,7 @@ Several bug fixes were made to YARN depl
 * Fixed bug in Python de-pickling 
[[SPARK-1135](https://issues.apache.org/jira/browse/SPARK-1135)]
 * Fixed bug in serialization of strings longer than 64K 
[[SPARK-1043](https://issues.apache.org/jira/browse/SPARK-1043)] 
 * Fixed bug that made jobs hang when base file is not available 
[[SPARK-1025](https://issues.apache.org/jira/browse/SPARK-1025)] 
-* Added Missing RDD operations to PySpark - top, zip, foldByKey, repartition, 
coallesce, getStorageLevel, setName and toDebugString
+* Added Missing RDD operations to PySpark - top, zip, foldByKey, repartition, 
coalesce, getStorageLevel, setName and toDebugString
 
 ### Improvements to documentation
 * Streaming: Added documentation on running streaming application from 
spark-shell

Modified: spark/site/releases/spark-release-0-9-1.html
URL: 
http://svn.apache.org/viewvc/spark/site/releases/spark-release-0-9-1.html?rev=1586197r1=1586196r2=1586197view=diff
==
--- spark/site/releases/spark-release-0-9-1.html (original)
+++ spark/site/releases/spark-release-0-9-1.html Thu Apr 10 02:17:44 2014
@@ -217,7 +217,7 @@
   liFixed bug in Python de-pickling [a 
href=https://issues.apache.org/jira/browse/SPARK-1135;SPARK-1135/a]/li
   liFixed bug in serialization of strings longer than 64K [a 
href=https://issues.apache.org/jira/browse/SPARK-1043;SPARK-1043/a] /li
   liFixed bug that made jobs hang when base file is not available [a 
href=https://issues.apache.org/jira/browse/SPARK-1025;SPARK-1025/a] /li
-  liAdded Missing RDD operations to PySpark - top, zip, foldByKey, 
repartition, coallesce, getStorageLevel, setName and toDebugString/li
+  liAdded Missing RDD operations to PySpark - top, zip, foldByKey, 
repartition, coalesce, getStorageLevel, setName and toDebugString/li
 /ul
 
 h3 id=improvements-to-documentationImprovements to documentation/h3




git commit: [SPARK-1592][streaming] Automatically remove streaming input blocks

2014-04-24 Thread tdas
Repository: spark
Updated Branches:
  refs/heads/master 35e3d199f - 526a518bf


[SPARK-1592][streaming] Automatically remove streaming input blocks

The raw input data is stored as blocks in BlockManagers. Earlier they were 
cleared by cleaner ttl. Now since streaming does not require cleaner TTL to be 
set, the block would not get cleared. This increases up the Spark's memory 
usage, which is not even accounted and shown in the Spark storage UI. It may 
cause the data blocks to spill over to disk, which eventually slows down the 
receiving of data (persisting to memory become bottlenecked by writing to disk).

The solution in this PR is to automatically remove those blocks. The mechanism 
to keep track of which BlockRDDs (which has presents the raw data blocks as a 
RDD) can be safely cleared already exists. Just use it to explicitly remove 
blocks from BlockRDDs.

Author: Tathagata Das tathagata.das1...@gmail.com

Closes #512 from tdas/block-rdd-unpersist and squashes the following commits:

d25e610 [Tathagata Das] Merge remote-tracking branch 'apache/master' into 
block-rdd-unpersist
5f46d69 [Tathagata Das] Merge remote-tracking branch 'apache/master' into 
block-rdd-unpersist
2c320cd [Tathagata Das] Updated configuration with spark.streaming.unpersist 
setting.
2d4b2fd [Tathagata Das] Automatically removed input blocks


Project: http://git-wip-us.apache.org/repos/asf/spark/repo
Commit: http://git-wip-us.apache.org/repos/asf/spark/commit/526a518b
Tree: http://git-wip-us.apache.org/repos/asf/spark/tree/526a518b
Diff: http://git-wip-us.apache.org/repos/asf/spark/diff/526a518b

Branch: refs/heads/master
Commit: 526a518bf32ad55b926a26f16086f445fd0ae29f
Parents: 35e3d19
Author: Tathagata Das tathagata.das1...@gmail.com
Authored: Thu Apr 24 18:18:22 2014 -0700
Committer: Tathagata Das tathagata.das1...@gmail.com
Committed: Thu Apr 24 18:18:22 2014 -0700

--
 .../scala/org/apache/spark/rdd/BlockRDD.scala   | 45 ++--
 docs/configuration.md   |  7 +-
 .../scala/org/apache/spark/streaming/Time.scala |  2 +-
 .../spark/streaming/dstream/DStream.scala   | 16 -
 .../spark/streaming/BasicOperationsSuite.scala  | 76 +++-
 .../spark/streaming/InputStreamsSuite.scala | 13 
 .../spark/streaming/NetworkReceiverSuite.scala  |  1 +
 7 files changed, 135 insertions(+), 25 deletions(-)
--


http://git-wip-us.apache.org/repos/asf/spark/blob/526a518b/core/src/main/scala/org/apache/spark/rdd/BlockRDD.scala
--
diff --git a/core/src/main/scala/org/apache/spark/rdd/BlockRDD.scala 
b/core/src/main/scala/org/apache/spark/rdd/BlockRDD.scala
index e6c4a6d..c64da88 100644
--- a/core/src/main/scala/org/apache/spark/rdd/BlockRDD.scala
+++ b/core/src/main/scala/org/apache/spark/rdd/BlockRDD.scala
@@ -19,24 +19,30 @@ package org.apache.spark.rdd
 
 import scala.reflect.ClassTag
 
-import org.apache.spark.{Partition, SparkContext, SparkEnv, TaskContext}
+import org.apache.spark._
 import org.apache.spark.storage.{BlockId, BlockManager}
+import scala.Some
 
 private[spark] class BlockRDDPartition(val blockId: BlockId, idx: Int) extends 
Partition {
   val index = idx
 }
 
 private[spark]
-class BlockRDD[T: ClassTag](sc: SparkContext, @transient blockIds: 
Array[BlockId])
+class BlockRDD[T: ClassTag](@transient sc: SparkContext, @transient val 
blockIds: Array[BlockId])
   extends RDD[T](sc, Nil) {
 
   @transient lazy val locations_ = BlockManager.blockIdsToHosts(blockIds, 
SparkEnv.get)
+  @volatile private var _isValid = true
 
-  override def getPartitions: Array[Partition] = (0 until blockIds.size).map(i 
= {
-new BlockRDDPartition(blockIds(i), i).asInstanceOf[Partition]
-  }).toArray
+  override def getPartitions: Array[Partition] = {
+assertValid()
+(0 until blockIds.size).map(i = {
+  new BlockRDDPartition(blockIds(i), i).asInstanceOf[Partition]
+}).toArray
+  }
 
   override def compute(split: Partition, context: TaskContext): Iterator[T] = {
+assertValid()
 val blockManager = SparkEnv.get.blockManager
 val blockId = split.asInstanceOf[BlockRDDPartition].blockId
 blockManager.get(blockId) match {
@@ -47,7 +53,36 @@ class BlockRDD[T: ClassTag](sc: SparkContext, @transient 
blockIds: Array[BlockId
   }
 
   override def getPreferredLocations(split: Partition): Seq[String] = {
+assertValid()
 locations_(split.asInstanceOf[BlockRDDPartition].blockId)
   }
+
+  /**
+   * Remove the data blocks that this BlockRDD is made from. NOTE: This is an
+   * irreversible operation, as the data in the blocks cannot be recovered back
+   * once removed. Use it with caution.
+   */
+  private[spark] def removeBlocks() {
+blockIds.foreach { blockId =
+  sc.env.blockManager.master.removeBlock(blockId)
+}
+_isValid = false

git commit: [SPARK-1592][streaming] Automatically remove streaming input blocks

2014-04-24 Thread tdas
Repository: spark
Updated Branches:
  refs/heads/branch-1.0 521d43599 - a3b6d8523


[SPARK-1592][streaming] Automatically remove streaming input blocks

The raw input data is stored as blocks in BlockManagers. Earlier they were 
cleared by cleaner ttl. Now since streaming does not require cleaner TTL to be 
set, the block would not get cleared. This increases up the Spark's memory 
usage, which is not even accounted and shown in the Spark storage UI. It may 
cause the data blocks to spill over to disk, which eventually slows down the 
receiving of data (persisting to memory become bottlenecked by writing to disk).

The solution in this PR is to automatically remove those blocks. The mechanism 
to keep track of which BlockRDDs (which has presents the raw data blocks as a 
RDD) can be safely cleared already exists. Just use it to explicitly remove 
blocks from BlockRDDs.

Author: Tathagata Das tathagata.das1...@gmail.com

Closes #512 from tdas/block-rdd-unpersist and squashes the following commits:

d25e610 [Tathagata Das] Merge remote-tracking branch 'apache/master' into 
block-rdd-unpersist
5f46d69 [Tathagata Das] Merge remote-tracking branch 'apache/master' into 
block-rdd-unpersist
2c320cd [Tathagata Das] Updated configuration with spark.streaming.unpersist 
setting.
2d4b2fd [Tathagata Das] Automatically removed input blocks

(cherry picked from commit 526a518bf32ad55b926a26f16086f445fd0ae29f)
Signed-off-by: Tathagata Das tathagata.das1...@gmail.com


Project: http://git-wip-us.apache.org/repos/asf/spark/repo
Commit: http://git-wip-us.apache.org/repos/asf/spark/commit/a3b6d852
Tree: http://git-wip-us.apache.org/repos/asf/spark/tree/a3b6d852
Diff: http://git-wip-us.apache.org/repos/asf/spark/diff/a3b6d852

Branch: refs/heads/branch-1.0
Commit: a3b6d852337c5c4c9afd4942699f35baaa6f691d
Parents: 521d435
Author: Tathagata Das tathagata.das1...@gmail.com
Authored: Thu Apr 24 18:18:22 2014 -0700
Committer: Tathagata Das tathagata.das1...@gmail.com
Committed: Thu Apr 24 18:18:47 2014 -0700

--
 .../scala/org/apache/spark/rdd/BlockRDD.scala   | 45 ++--
 docs/configuration.md   |  7 +-
 .../scala/org/apache/spark/streaming/Time.scala |  2 +-
 .../spark/streaming/dstream/DStream.scala   | 16 -
 .../spark/streaming/BasicOperationsSuite.scala  | 76 +++-
 .../spark/streaming/InputStreamsSuite.scala | 13 
 .../spark/streaming/NetworkReceiverSuite.scala  |  1 +
 7 files changed, 135 insertions(+), 25 deletions(-)
--


http://git-wip-us.apache.org/repos/asf/spark/blob/a3b6d852/core/src/main/scala/org/apache/spark/rdd/BlockRDD.scala
--
diff --git a/core/src/main/scala/org/apache/spark/rdd/BlockRDD.scala 
b/core/src/main/scala/org/apache/spark/rdd/BlockRDD.scala
index e6c4a6d..c64da88 100644
--- a/core/src/main/scala/org/apache/spark/rdd/BlockRDD.scala
+++ b/core/src/main/scala/org/apache/spark/rdd/BlockRDD.scala
@@ -19,24 +19,30 @@ package org.apache.spark.rdd
 
 import scala.reflect.ClassTag
 
-import org.apache.spark.{Partition, SparkContext, SparkEnv, TaskContext}
+import org.apache.spark._
 import org.apache.spark.storage.{BlockId, BlockManager}
+import scala.Some
 
 private[spark] class BlockRDDPartition(val blockId: BlockId, idx: Int) extends 
Partition {
   val index = idx
 }
 
 private[spark]
-class BlockRDD[T: ClassTag](sc: SparkContext, @transient blockIds: 
Array[BlockId])
+class BlockRDD[T: ClassTag](@transient sc: SparkContext, @transient val 
blockIds: Array[BlockId])
   extends RDD[T](sc, Nil) {
 
   @transient lazy val locations_ = BlockManager.blockIdsToHosts(blockIds, 
SparkEnv.get)
+  @volatile private var _isValid = true
 
-  override def getPartitions: Array[Partition] = (0 until blockIds.size).map(i 
= {
-new BlockRDDPartition(blockIds(i), i).asInstanceOf[Partition]
-  }).toArray
+  override def getPartitions: Array[Partition] = {
+assertValid()
+(0 until blockIds.size).map(i = {
+  new BlockRDDPartition(blockIds(i), i).asInstanceOf[Partition]
+}).toArray
+  }
 
   override def compute(split: Partition, context: TaskContext): Iterator[T] = {
+assertValid()
 val blockManager = SparkEnv.get.blockManager
 val blockId = split.asInstanceOf[BlockRDDPartition].blockId
 blockManager.get(blockId) match {
@@ -47,7 +53,36 @@ class BlockRDD[T: ClassTag](sc: SparkContext, @transient 
blockIds: Array[BlockId
   }
 
   override def getPreferredLocations(split: Partition): Seq[String] = {
+assertValid()
 locations_(split.asInstanceOf[BlockRDDPartition].blockId)
   }
+
+  /**
+   * Remove the data blocks that this BlockRDD is made from. NOTE: This is an
+   * irreversible operation, as the data in the blocks cannot be recovered back
+   * once removed. Use it with caution.
+   */
+  private[spark] def

git commit: [SPARK-1510] Spark Streaming metrics source for metrics system

2014-04-24 Thread tdas
Repository: spark
Updated Branches:
  refs/heads/master 44da5ab2d - 80429f3e2


[SPARK-1510] Spark Streaming metrics source for metrics system

This pulls in changes made by @jerryshao in 
https://github.com/apache/spark/pull/424 and merges with the master.

Author: jerryshao saisai.s...@intel.com
Author: Tathagata Das tathagata.das1...@gmail.com

Closes #545 from tdas/streaming-metrics and squashes the following commits:

034b443 [Tathagata Das] Merge remote-tracking branch 'apache-github/master' 
into streaming-metrics
fb3b0a5 [jerryshao] Modify according master update
21939f5 [jerryshao] Style changes according to style check error
976116b [jerryshao] Add StreamSource in StreamingContext for better monitoring 
through metrics system


Project: http://git-wip-us.apache.org/repos/asf/spark/repo
Commit: http://git-wip-us.apache.org/repos/asf/spark/commit/80429f3e
Tree: http://git-wip-us.apache.org/repos/asf/spark/tree/80429f3e
Diff: http://git-wip-us.apache.org/repos/asf/spark/diff/80429f3e

Branch: refs/heads/master
Commit: 80429f3e2ab786d103297652922c3d8da3cf5a01
Parents: 44da5ab
Author: jerryshao saisai.s...@intel.com
Authored: Thu Apr 24 18:56:57 2014 -0700
Committer: Tathagata Das tathagata.das1...@gmail.com
Committed: Thu Apr 24 18:56:57 2014 -0700

--
 .../spark/streaming/StreamingContext.scala  |  4 ++
 .../spark/streaming/StreamingSource.scala   | 73 
 .../ui/StreamingJobProgressListener.scala   |  3 +-
 3 files changed, 79 insertions(+), 1 deletion(-)
--


http://git-wip-us.apache.org/repos/asf/spark/blob/80429f3e/streaming/src/main/scala/org/apache/spark/streaming/StreamingContext.scala
--
diff --git 
a/streaming/src/main/scala/org/apache/spark/streaming/StreamingContext.scala 
b/streaming/src/main/scala/org/apache/spark/streaming/StreamingContext.scala
index 1c89543..e0677b7 100644
--- a/streaming/src/main/scala/org/apache/spark/streaming/StreamingContext.scala
+++ b/streaming/src/main/scala/org/apache/spark/streaming/StreamingContext.scala
@@ -154,6 +154,10 @@ class StreamingContext private[streaming] (
 
   private[streaming] val uiTab = new StreamingTab(this)
 
+  /** Register streaming source to metrics system */
+  private val streamingSource = new StreamingSource(this)
+  SparkEnv.get.metricsSystem.registerSource(streamingSource)
+
   /** Enumeration to identify current state of the StreamingContext */
   private[streaming] object StreamingContextState extends Enumeration {
 type CheckpointState = Value

http://git-wip-us.apache.org/repos/asf/spark/blob/80429f3e/streaming/src/main/scala/org/apache/spark/streaming/StreamingSource.scala
--
diff --git 
a/streaming/src/main/scala/org/apache/spark/streaming/StreamingSource.scala 
b/streaming/src/main/scala/org/apache/spark/streaming/StreamingSource.scala
new file mode 100644
index 000..774adc3
--- /dev/null
+++ b/streaming/src/main/scala/org/apache/spark/streaming/StreamingSource.scala
@@ -0,0 +1,73 @@
+/*
+ * Licensed to the Apache Software Foundation (ASF) under one or more
+ * contributor license agreements.  See the NOTICE file distributed with
+ * this work for additional information regarding copyright ownership.
+ * The ASF licenses this file to You under the Apache License, Version 2.0
+ * (the License); you may not use this file except in compliance with
+ * the License.  You may obtain a copy of the License at
+ *
+ *http://www.apache.org/licenses/LICENSE-2.0
+ *
+ * Unless required by applicable law or agreed to in writing, software
+ * distributed under the License is distributed on an AS IS BASIS,
+ * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+ * See the License for the specific language governing permissions and
+ * limitations under the License.
+ */
+
+package org.apache.spark.streaming
+
+import com.codahale.metrics.{Gauge, MetricRegistry}
+
+import org.apache.spark.metrics.source.Source
+import org.apache.spark.streaming.ui.StreamingJobProgressListener
+
+private[streaming] class StreamingSource(ssc: StreamingContext) extends Source 
{
+  val metricRegistry = new MetricRegistry
+  val sourceName = %s.StreamingMetrics.format(ssc.sparkContext.appName)
+
+  val streamingListener = ssc.uiTab.listener
+
+  private def registerGauge[T](name: String, f: StreamingJobProgressListener 
= T,
+  defaultValue: T) {
+metricRegistry.register(MetricRegistry.name(streaming, name), new 
Gauge[T] {
+  override def getValue: T = 
Option(f(streamingListener)).getOrElse(defaultValue)
+})
+  }
+
+  // Gauge for number of network receivers
+  registerGauge(receivers, _.numReceivers, 0)
+
+  // Gauge for number of total completed batches
+  registerGauge(totalCompletedBatches

git commit: [SPARK-1510] Spark Streaming metrics source for metrics system

2014-04-24 Thread tdas
Repository: spark
Updated Branches:
  refs/heads/branch-1.0 c8dd13221 - 0bc0f36d6


[SPARK-1510] Spark Streaming metrics source for metrics system

This pulls in changes made by @jerryshao in 
https://github.com/apache/spark/pull/424 and merges with the master.

Author: jerryshao saisai.s...@intel.com
Author: Tathagata Das tathagata.das1...@gmail.com

Closes #545 from tdas/streaming-metrics and squashes the following commits:

034b443 [Tathagata Das] Merge remote-tracking branch 'apache-github/master' 
into streaming-metrics
fb3b0a5 [jerryshao] Modify according master update
21939f5 [jerryshao] Style changes according to style check error
976116b [jerryshao] Add StreamSource in StreamingContext for better monitoring 
through metrics system

(cherry picked from commit 80429f3e2ab786d103297652922c3d8da3cf5a01)
Signed-off-by: Tathagata Das tathagata.das1...@gmail.com


Project: http://git-wip-us.apache.org/repos/asf/spark/repo
Commit: http://git-wip-us.apache.org/repos/asf/spark/commit/0bc0f36d
Tree: http://git-wip-us.apache.org/repos/asf/spark/tree/0bc0f36d
Diff: http://git-wip-us.apache.org/repos/asf/spark/diff/0bc0f36d

Branch: refs/heads/branch-1.0
Commit: 0bc0f36d6be265d8ad5b1909fc06f03a661454a5
Parents: c8dd132
Author: jerryshao saisai.s...@intel.com
Authored: Thu Apr 24 18:56:57 2014 -0700
Committer: Tathagata Das tathagata.das1...@gmail.com
Committed: Thu Apr 24 18:57:11 2014 -0700

--
 .../spark/streaming/StreamingContext.scala  |  4 ++
 .../spark/streaming/StreamingSource.scala   | 73 
 .../ui/StreamingJobProgressListener.scala   |  3 +-
 3 files changed, 79 insertions(+), 1 deletion(-)
--


http://git-wip-us.apache.org/repos/asf/spark/blob/0bc0f36d/streaming/src/main/scala/org/apache/spark/streaming/StreamingContext.scala
--
diff --git 
a/streaming/src/main/scala/org/apache/spark/streaming/StreamingContext.scala 
b/streaming/src/main/scala/org/apache/spark/streaming/StreamingContext.scala
index 1c89543..e0677b7 100644
--- a/streaming/src/main/scala/org/apache/spark/streaming/StreamingContext.scala
+++ b/streaming/src/main/scala/org/apache/spark/streaming/StreamingContext.scala
@@ -154,6 +154,10 @@ class StreamingContext private[streaming] (
 
   private[streaming] val uiTab = new StreamingTab(this)
 
+  /** Register streaming source to metrics system */
+  private val streamingSource = new StreamingSource(this)
+  SparkEnv.get.metricsSystem.registerSource(streamingSource)
+
   /** Enumeration to identify current state of the StreamingContext */
   private[streaming] object StreamingContextState extends Enumeration {
 type CheckpointState = Value

http://git-wip-us.apache.org/repos/asf/spark/blob/0bc0f36d/streaming/src/main/scala/org/apache/spark/streaming/StreamingSource.scala
--
diff --git 
a/streaming/src/main/scala/org/apache/spark/streaming/StreamingSource.scala 
b/streaming/src/main/scala/org/apache/spark/streaming/StreamingSource.scala
new file mode 100644
index 000..774adc3
--- /dev/null
+++ b/streaming/src/main/scala/org/apache/spark/streaming/StreamingSource.scala
@@ -0,0 +1,73 @@
+/*
+ * Licensed to the Apache Software Foundation (ASF) under one or more
+ * contributor license agreements.  See the NOTICE file distributed with
+ * this work for additional information regarding copyright ownership.
+ * The ASF licenses this file to You under the Apache License, Version 2.0
+ * (the License); you may not use this file except in compliance with
+ * the License.  You may obtain a copy of the License at
+ *
+ *http://www.apache.org/licenses/LICENSE-2.0
+ *
+ * Unless required by applicable law or agreed to in writing, software
+ * distributed under the License is distributed on an AS IS BASIS,
+ * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+ * See the License for the specific language governing permissions and
+ * limitations under the License.
+ */
+
+package org.apache.spark.streaming
+
+import com.codahale.metrics.{Gauge, MetricRegistry}
+
+import org.apache.spark.metrics.source.Source
+import org.apache.spark.streaming.ui.StreamingJobProgressListener
+
+private[streaming] class StreamingSource(ssc: StreamingContext) extends Source 
{
+  val metricRegistry = new MetricRegistry
+  val sourceName = %s.StreamingMetrics.format(ssc.sparkContext.appName)
+
+  val streamingListener = ssc.uiTab.listener
+
+  private def registerGauge[T](name: String, f: StreamingJobProgressListener 
= T,
+  defaultValue: T) {
+metricRegistry.register(MetricRegistry.name(streaming, name), new 
Gauge[T] {
+  override def getValue: T = 
Option(f(streamingListener)).getOrElse(defaultValue)
+})
+  }
+
+  // Gauge for number of network receivers
+  registerGauge

git commit: [SPARK-1617] and [SPARK-1618] Improvements to streaming ui and bug fix to socket receiver

2014-04-24 Thread tdas
Repository: spark
Updated Branches:
  refs/heads/master 968c0187a - cd12dd9bd


[SPARK-1617] and [SPARK-1618] Improvements to streaming ui and bug fix to 
socket receiver

1617: These changes expose the receiver state (active or inactive) and last 
error in the UI
1618: If the socket receiver cannot connect in the first attempt, it should try 
to restart after a delay. That was broken, as the thread that restarts (hence, 
stops) the receiver waited on Thread.join on itself!

Author: Tathagata Das tathagata.das1...@gmail.com

Closes #540 from tdas/streaming-ui-fix and squashes the following commits:

e469434 [Tathagata Das] Merge remote-tracking branch 'apache-github/master' 
into streaming-ui-fix
dbddf75 [Tathagata Das] Style fix.
66df1a5 [Tathagata Das] Merge remote-tracking branch 'apache/master' into 
streaming-ui-fix
ad98bc9 [Tathagata Das] Refactored streaming listener to use ReceiverInfo.
d7f849c [Tathagata Das] Revert Moved BatchInfo from streaming.scheduler to 
streaming.ui
5c80919 [Tathagata Das] Moved BatchInfo from streaming.scheduler to streaming.ui
da244f6 [Tathagata Das] Fixed socket receiver as well as made receiver state 
and error visible in the streamign UI.


Project: http://git-wip-us.apache.org/repos/asf/spark/repo
Commit: http://git-wip-us.apache.org/repos/asf/spark/commit/cd12dd9b
Tree: http://git-wip-us.apache.org/repos/asf/spark/tree/cd12dd9b
Diff: http://git-wip-us.apache.org/repos/asf/spark/diff/cd12dd9b

Branch: refs/heads/master
Commit: cd12dd9bde91303d0341180e5f70d2a03d6b65b6
Parents: 968c018
Author: Tathagata Das tathagata.das1...@gmail.com
Authored: Thu Apr 24 21:34:37 2014 -0700
Committer: Tathagata Das tathagata.das1...@gmail.com
Committed: Thu Apr 24 21:34:37 2014 -0700

--
 .../scala/org/apache/spark/ui/UIUtils.scala |  2 +-
 .../streaming/dstream/SocketInputDStream.scala  | 49 +++--
 .../streaming/receiver/ActorReceiver.scala  | 12 +++-
 .../spark/streaming/receiver/Receiver.scala |  5 +-
 .../streaming/receiver/ReceiverMessage.scala|  4 +-
 .../streaming/receiver/ReceiverSupervisor.scala | 58 
 .../receiver/ReceiverSupervisorImpl.scala   | 24 +---
 .../spark/streaming/scheduler/BatchInfo.scala   |  3 +
 .../streaming/scheduler/ReceiverInfo.scala  | 37 +
 .../streaming/scheduler/ReceiverTracker.scala   | 40 +-
 .../streaming/scheduler/StreamingListener.scala | 25 -
 .../ui/StreamingJobProgressListener.scala   | 18 +-
 .../spark/streaming/ui/StreamingPage.scala  | 20 +--
 .../spark/streaming/NetworkReceiverSuite.scala  |  8 ++-
 .../streaming/StreamingListenerSuite.scala  | 15 ++---
 15 files changed, 217 insertions(+), 103 deletions(-)
--


http://git-wip-us.apache.org/repos/asf/spark/blob/cd12dd9b/core/src/main/scala/org/apache/spark/ui/UIUtils.scala
--
diff --git a/core/src/main/scala/org/apache/spark/ui/UIUtils.scala 
b/core/src/main/scala/org/apache/spark/ui/UIUtils.scala
index cf987a1..a3d6a18 100644
--- a/core/src/main/scala/org/apache/spark/ui/UIUtils.scala
+++ b/core/src/main/scala/org/apache/spark/ui/UIUtils.scala
@@ -122,7 +122,7 @@ private[spark] object UIUtils extends Logging {
   }
 }
 if (unit.isEmpty) {
-  %d.formatLocal(Locale.US, value)
+  %d.formatLocal(Locale.US, value.toInt)
 } else {
   %.1f%s.formatLocal(Locale.US, value, unit)
 }

http://git-wip-us.apache.org/repos/asf/spark/blob/cd12dd9b/streaming/src/main/scala/org/apache/spark/streaming/dstream/SocketInputDStream.scala
--
diff --git 
a/streaming/src/main/scala/org/apache/spark/streaming/dstream/SocketInputDStream.scala
 
b/streaming/src/main/scala/org/apache/spark/streaming/dstream/SocketInputDStream.scala
index 1e32727..8b72bcf 100644
--- 
a/streaming/src/main/scala/org/apache/spark/streaming/dstream/SocketInputDStream.scala
+++ 
b/streaming/src/main/scala/org/apache/spark/streaming/dstream/SocketInputDStream.scala
@@ -50,49 +50,42 @@ class SocketReceiver[T: ClassTag](
 storageLevel: StorageLevel
   ) extends Receiver[T](storageLevel) with Logging {
 
-  var socket: Socket = null
-  var receivingThread: Thread = null
-
   def onStart() {
-receivingThread = new Thread(Socket Receiver) {
-  override def run() {
-connect()
-receive()
-  }
-}
-receivingThread.start()
+// Start the thread that receives data over a connection
+new Thread(Socket Receiver) {
+  setDaemon(true)
+  override def run() { receive() }
+}.start()
   }
 
   def onStop() {
-if (socket != null) {
-  socket.close()
-}
-socket = null
-if (receivingThread != null) {
-  receivingThread.join()
-}
+// There is nothing much to do

git commit: [SPARK-1617] and [SPARK-1618] Improvements to streaming ui and bug fix to socket receiver

2014-04-24 Thread tdas
Repository: spark
Updated Branches:
  refs/heads/branch-1.0 777a9a5a1 - d933c710f


[SPARK-1617] and [SPARK-1618] Improvements to streaming ui and bug fix to 
socket receiver

1617: These changes expose the receiver state (active or inactive) and last 
error in the UI
1618: If the socket receiver cannot connect in the first attempt, it should try 
to restart after a delay. That was broken, as the thread that restarts (hence, 
stops) the receiver waited on Thread.join on itself!

Author: Tathagata Das tathagata.das1...@gmail.com

Closes #540 from tdas/streaming-ui-fix and squashes the following commits:

e469434 [Tathagata Das] Merge remote-tracking branch 'apache-github/master' 
into streaming-ui-fix
dbddf75 [Tathagata Das] Style fix.
66df1a5 [Tathagata Das] Merge remote-tracking branch 'apache/master' into 
streaming-ui-fix
ad98bc9 [Tathagata Das] Refactored streaming listener to use ReceiverInfo.
d7f849c [Tathagata Das] Revert Moved BatchInfo from streaming.scheduler to 
streaming.ui
5c80919 [Tathagata Das] Moved BatchInfo from streaming.scheduler to streaming.ui
da244f6 [Tathagata Das] Fixed socket receiver as well as made receiver state 
and error visible in the streamign UI.

(cherry picked from commit cd12dd9bde91303d0341180e5f70d2a03d6b65b6)
Signed-off-by: Tathagata Das tathagata.das1...@gmail.com


Project: http://git-wip-us.apache.org/repos/asf/spark/repo
Commit: http://git-wip-us.apache.org/repos/asf/spark/commit/d933c710
Tree: http://git-wip-us.apache.org/repos/asf/spark/tree/d933c710
Diff: http://git-wip-us.apache.org/repos/asf/spark/diff/d933c710

Branch: refs/heads/branch-1.0
Commit: d933c710f9b49f398446d51859fe6953d2e4b6f9
Parents: 777a9a5
Author: Tathagata Das tathagata.das1...@gmail.com
Authored: Thu Apr 24 21:34:37 2014 -0700
Committer: Tathagata Das tathagata.das1...@gmail.com
Committed: Thu Apr 24 21:34:50 2014 -0700

--
 .../scala/org/apache/spark/ui/UIUtils.scala |  2 +-
 .../streaming/dstream/SocketInputDStream.scala  | 49 +++--
 .../streaming/receiver/ActorReceiver.scala  | 12 +++-
 .../spark/streaming/receiver/Receiver.scala |  5 +-
 .../streaming/receiver/ReceiverMessage.scala|  4 +-
 .../streaming/receiver/ReceiverSupervisor.scala | 58 
 .../receiver/ReceiverSupervisorImpl.scala   | 24 +---
 .../spark/streaming/scheduler/BatchInfo.scala   |  3 +
 .../streaming/scheduler/ReceiverInfo.scala  | 37 +
 .../streaming/scheduler/ReceiverTracker.scala   | 40 +-
 .../streaming/scheduler/StreamingListener.scala | 25 -
 .../ui/StreamingJobProgressListener.scala   | 18 +-
 .../spark/streaming/ui/StreamingPage.scala  | 20 +--
 .../spark/streaming/NetworkReceiverSuite.scala  |  8 ++-
 .../streaming/StreamingListenerSuite.scala  | 15 ++---
 15 files changed, 217 insertions(+), 103 deletions(-)
--


http://git-wip-us.apache.org/repos/asf/spark/blob/d933c710/core/src/main/scala/org/apache/spark/ui/UIUtils.scala
--
diff --git a/core/src/main/scala/org/apache/spark/ui/UIUtils.scala 
b/core/src/main/scala/org/apache/spark/ui/UIUtils.scala
index cf987a1..a3d6a18 100644
--- a/core/src/main/scala/org/apache/spark/ui/UIUtils.scala
+++ b/core/src/main/scala/org/apache/spark/ui/UIUtils.scala
@@ -122,7 +122,7 @@ private[spark] object UIUtils extends Logging {
   }
 }
 if (unit.isEmpty) {
-  %d.formatLocal(Locale.US, value)
+  %d.formatLocal(Locale.US, value.toInt)
 } else {
   %.1f%s.formatLocal(Locale.US, value, unit)
 }

http://git-wip-us.apache.org/repos/asf/spark/blob/d933c710/streaming/src/main/scala/org/apache/spark/streaming/dstream/SocketInputDStream.scala
--
diff --git 
a/streaming/src/main/scala/org/apache/spark/streaming/dstream/SocketInputDStream.scala
 
b/streaming/src/main/scala/org/apache/spark/streaming/dstream/SocketInputDStream.scala
index 1e32727..8b72bcf 100644
--- 
a/streaming/src/main/scala/org/apache/spark/streaming/dstream/SocketInputDStream.scala
+++ 
b/streaming/src/main/scala/org/apache/spark/streaming/dstream/SocketInputDStream.scala
@@ -50,49 +50,42 @@ class SocketReceiver[T: ClassTag](
 storageLevel: StorageLevel
   ) extends Receiver[T](storageLevel) with Logging {
 
-  var socket: Socket = null
-  var receivingThread: Thread = null
-
   def onStart() {
-receivingThread = new Thread(Socket Receiver) {
-  override def run() {
-connect()
-receive()
-  }
-}
-receivingThread.start()
+// Start the thread that receives data over a connection
+new Thread(Socket Receiver) {
+  setDaemon(true)
+  override def run() { receive() }
+}.start()
   }
 
   def onStop() {
-if (socket != null) {
-  socket.close

git commit: [Spark-1382] Fix NPE in DStream.slice (updated version of #365)

2014-04-25 Thread tdas
Repository: spark
Updated Branches:
  refs/heads/branch-1.0 94c71e0a2 - f85c6815e


[Spark-1382] Fix NPE in DStream.slice (updated version of #365)

@zsxwing I cherry-picked your changes and merged the master. #365 had some 
conflicts once again!

Author: zsxwing zsxw...@gmail.com
Author: Tathagata Das tathagata.das1...@gmail.com

Closes #562 from tdas/SPARK-1382 and squashes the following commits:

e2962c1 [Tathagata Das] Merge remote-tracking branch 'apache-github/master' 
into SPARK-1382
20968d9 [zsxwing] Replace Exception with SparkException in DStream
e476651 [zsxwing] Merge remote-tracking branch 'origin/master' into SPARK-1382
35ba56a [zsxwing] SPARK-1382: Fix NPE in DStream.slice

(cherry picked from commit 058797c1722c9251f6bc6ad2672cb0e79146b04f)
Signed-off-by: Tathagata Das tathagata.das1...@gmail.com


Project: http://git-wip-us.apache.org/repos/asf/spark/repo
Commit: http://git-wip-us.apache.org/repos/asf/spark/commit/f85c6815
Tree: http://git-wip-us.apache.org/repos/asf/spark/tree/f85c6815
Diff: http://git-wip-us.apache.org/repos/asf/spark/diff/f85c6815

Branch: refs/heads/branch-1.0
Commit: f85c6815ebb80cdc35e63c7e3c2a354a808e4249
Parents: 94c71e0
Author: zsxwing zsxw...@gmail.com
Authored: Fri Apr 25 19:04:34 2014 -0700
Committer: Tathagata Das tathagata.das1...@gmail.com
Committed: Fri Apr 25 19:04:45 2014 -0700

--
 .../spark/streaming/dstream/DStream.scala   | 22 +++-
 .../spark/streaming/BasicOperationsSuite.scala  | 12 ++-
 2 files changed, 23 insertions(+), 11 deletions(-)
--


http://git-wip-us.apache.org/repos/asf/spark/blob/f85c6815/streaming/src/main/scala/org/apache/spark/streaming/dstream/DStream.scala
--
diff --git 
a/streaming/src/main/scala/org/apache/spark/streaming/dstream/DStream.scala 
b/streaming/src/main/scala/org/apache/spark/streaming/dstream/DStream.scala
index f69f69e..4709a62 100644
--- a/streaming/src/main/scala/org/apache/spark/streaming/dstream/DStream.scala
+++ b/streaming/src/main/scala/org/apache/spark/streaming/dstream/DStream.scala
@@ -18,20 +18,19 @@
 package org.apache.spark.streaming.dstream
 
 
+import java.io.{IOException, ObjectInputStream, ObjectOutputStream}
+
 import scala.deprecated
 import scala.collection.mutable.HashMap
 import scala.reflect.ClassTag
 
-import java.io.{IOException, ObjectInputStream, ObjectOutputStream}
-
-import org.apache.spark.Logging
+import org.apache.spark.{Logging, SparkException}
 import org.apache.spark.rdd.{BlockRDD, RDD}
 import org.apache.spark.storage.StorageLevel
-import org.apache.spark.util.MetadataCleaner
 import org.apache.spark.streaming._
 import org.apache.spark.streaming.StreamingContext._
 import org.apache.spark.streaming.scheduler.Job
-import org.apache.spark.streaming.Duration
+import org.apache.spark.util.MetadataCleaner
 
 /**
  * A Discretized Stream (DStream), the basic abstraction in Spark Streaming, 
is a continuous
@@ -144,7 +143,7 @@ abstract class DStream[T: ClassTag] (
*/
   private[streaming] def initialize(time: Time) {
 if (zeroTime != null  zeroTime != time) {
-  throw new Exception(ZeroTime is already initialized to  + zeroTime
+  throw new SparkException(ZeroTime is already initialized to  + zeroTime
 + , cannot initialize it again to  + time)
 }
 zeroTime = time
@@ -220,7 +219,7 @@ abstract class DStream[T: ClassTag] (
 which requires  + this.getClass.getSimpleName +  to remember 
generated RDDs for more  +
 than  + rememberDuration.milliseconds / 1000 +  seconds. But 
Spark's metadata cleanup +
 delay is set to  + metadataCleanerDelay +  seconds, which is not 
sufficient. Please  +
-set the Java property 'spark.cleaner.delay' to more than  +
+set the Java cleaner delay to more than  +
 math.ceil(rememberDuration.milliseconds / 1000.0).toInt +  seconds.
 )
 
@@ -235,7 +234,7 @@ abstract class DStream[T: ClassTag] (
 
   private[streaming] def setContext(s: StreamingContext) {
 if (ssc != null  ssc != s) {
-  throw new Exception(Context is already set in  + this + , cannot set 
it again)
+  throw new SparkException(Context is already set in  + this + , cannot 
set it again)
 }
 ssc = s
 logInfo(Set context for  + this)
@@ -244,7 +243,7 @@ abstract class DStream[T: ClassTag] (
 
   private[streaming] def setGraph(g: DStreamGraph) {
 if (graph != null  graph != g) {
-  throw new Exception(Graph is already set in  + this + , cannot set it 
again)
+  throw new SparkException(Graph is already set in  + this + , cannot 
set it again)
 }
 graph = g
 dependencies.foreach(_.setGraph(graph))
@@ -261,7 +260,7 @@ abstract class DStream[T: ClassTag] (
   /** Checks whether the 'time' is valid wrt slideDuration for generating RDD

git commit: [SPARK-1504], [SPARK-1505], [SPARK-1558] Updated Spark Streaming guide

2014-05-05 Thread tdas
Repository: spark
Updated Branches:
  refs/heads/master 3292e2a71 - a975a19f2


[SPARK-1504], [SPARK-1505], [SPARK-1558] Updated Spark Streaming guide

- SPARK-1558: Updated custom receiver guide to match it with the new API
- SPARK-1504: Added deployment and monitoring subsection to streaming
- SPARK-1505: Added migration guide for migrating from 0.9.x and below to Spark 
1.0
- Updated various Java streaming examples to use JavaReceiverInputDStream to 
highlight the API change.
- Removed the requirement for cleaner ttl from streaming guide

Author: Tathagata Das tathagata.das1...@gmail.com

Closes #652 from tdas/doc-fix and squashes the following commits:

cb4f4b7 [Tathagata Das] Possible fix for flaky graceful shutdown test.
ab71f7f [Tathagata Das] Merge remote-tracking branch 'apache-github/master' 
into doc-fix
8d6ff9b [Tathagata Das] Addded migration guide to Spark Streaming.
7d171df [Tathagata Das] Added reference to JavaReceiverInputStream in examples 
and streaming guide.
49edd7c [Tathagata Das] Change java doc links to use Java docs.
11528d7 [Tathagata Das] Updated links on index page.
ff80970 [Tathagata Das] More updates to streaming guide.
4dc42e9 [Tathagata Das] Added monitoring and other documentation in the 
streaming guide.
14c6564 [Tathagata Das] Updated custom receiver guide.


Project: http://git-wip-us.apache.org/repos/asf/spark/repo
Commit: http://git-wip-us.apache.org/repos/asf/spark/commit/a975a19f
Tree: http://git-wip-us.apache.org/repos/asf/spark/tree/a975a19f
Diff: http://git-wip-us.apache.org/repos/asf/spark/diff/a975a19f

Branch: refs/heads/master
Commit: a975a19f21e71f448b3fdb2ed4461e28ef439900
Parents: 3292e2a
Author: Tathagata Das tathagata.das1...@gmail.com
Authored: Mon May 5 15:28:19 2014 -0700
Committer: Tathagata Das tathagata.das1...@gmail.com
Committed: Mon May 5 15:28:19 2014 -0700

--
 docs/configuration.md   |   6 +-
 docs/index.md   |   6 +-
 docs/streaming-custom-receivers.md  | 273 +--
 docs/streaming-programming-guide.md | 200 +++---
 .../streaming/examples/JavaCustomReceiver.java  |   3 +-
 .../streaming/examples/JavaFlumeEventCount.java |   2 +-
 .../streaming/examples/JavaKafkaWordCount.java  |   4 +-
 .../examples/JavaNetworkWordCount.java  |   5 +-
 .../spark/streaming/receiver/Receiver.scala |   6 +-
 .../spark/streaming/StreamingContextSuite.scala |   1 +
 10 files changed, 360 insertions(+), 146 deletions(-)
--


http://git-wip-us.apache.org/repos/asf/spark/blob/a975a19f/docs/configuration.md
--
diff --git a/docs/configuration.md b/docs/configuration.md
index 81ad895..d6f316b 100644
--- a/docs/configuration.md
+++ b/docs/configuration.md
@@ -462,7 +462,7 @@ Apart from these, the following properties are also 
available, and may be useful
   td(infinite)/td
   td
 Duration (seconds) of how long Spark will remember any metadata (stages 
generated, tasks generated, etc.).
-Periodic cleanups will ensure that metadata older than this duration will 
be forgetten. This is
+Periodic cleanups will ensure that metadata older than this duration will 
be forgotten. This is
 useful for running Spark for many hours / days (for example, running 24/7 
in case of Spark Streaming
 applications). Note that any RDD that persists in memory for more than 
this duration will be cleared as well.
   /td
@@ -471,8 +471,8 @@ Apart from these, the following properties are also 
available, and may be useful
   tdspark.streaming.blockInterval/td
   td200/td
   td
-Duration (milliseconds) of how long to batch new objects coming from 
network receivers used
-in Spark Streaming.
+Interval (milliseconds) at which data received by Spark Streaming 
receivers is coalesced
+into blocks of data before storing them in Spark.
   /td
 /tr
 tr

http://git-wip-us.apache.org/repos/asf/spark/blob/a975a19f/docs/index.md
--
diff --git a/docs/index.md b/docs/index.md
index 6fc9a4f..2daa208 100644
--- a/docs/index.md
+++ b/docs/index.md
@@ -112,10 +112,10 @@ Note that on Windows, you need to set the environment 
variables on separate line
 * [Shark](http://shark.cs.berkeley.edu): Apache Hive over Spark
 * [Mailing Lists](http://spark.apache.org/mailing-lists.html): ask questions 
about Spark here
 * [AMP Camps](http://ampcamp.berkeley.edu/): a series of training camps at UC 
Berkeley that featured talks and
-  exercises about Spark, Shark, Mesos, and more. 
[Videos](http://ampcamp.berkeley.edu/agenda-2012),
-  [slides](http://ampcamp.berkeley.edu/agenda-2012) and 
[exercises](http://ampcamp.berkeley.edu/exercises-2012) are
+  exercises about Spark, Shark, Spark Streaming, Mesos, and more

git commit: [SPARK-1504], [SPARK-1505], [SPARK-1558] Updated Spark Streaming guide

2014-05-05 Thread tdas
Repository: spark
Updated Branches:
  refs/heads/branch-1.0 80f4360e7 - 1fac4ecbd


[SPARK-1504], [SPARK-1505], [SPARK-1558] Updated Spark Streaming guide

- SPARK-1558: Updated custom receiver guide to match it with the new API
- SPARK-1504: Added deployment and monitoring subsection to streaming
- SPARK-1505: Added migration guide for migrating from 0.9.x and below to Spark 
1.0
- Updated various Java streaming examples to use JavaReceiverInputDStream to 
highlight the API change.
- Removed the requirement for cleaner ttl from streaming guide

Author: Tathagata Das tathagata.das1...@gmail.com

Closes #652 from tdas/doc-fix and squashes the following commits:

cb4f4b7 [Tathagata Das] Possible fix for flaky graceful shutdown test.
ab71f7f [Tathagata Das] Merge remote-tracking branch 'apache-github/master' 
into doc-fix
8d6ff9b [Tathagata Das] Addded migration guide to Spark Streaming.
7d171df [Tathagata Das] Added reference to JavaReceiverInputStream in examples 
and streaming guide.
49edd7c [Tathagata Das] Change java doc links to use Java docs.
11528d7 [Tathagata Das] Updated links on index page.
ff80970 [Tathagata Das] More updates to streaming guide.
4dc42e9 [Tathagata Das] Added monitoring and other documentation in the 
streaming guide.
14c6564 [Tathagata Das] Updated custom receiver guide.

(cherry picked from commit a975a19f21e71f448b3fdb2ed4461e28ef439900)
Signed-off-by: Tathagata Das tathagata.das1...@gmail.com


Project: http://git-wip-us.apache.org/repos/asf/spark/repo
Commit: http://git-wip-us.apache.org/repos/asf/spark/commit/1fac4ecb
Tree: http://git-wip-us.apache.org/repos/asf/spark/tree/1fac4ecb
Diff: http://git-wip-us.apache.org/repos/asf/spark/diff/1fac4ecb

Branch: refs/heads/branch-1.0
Commit: 1fac4ecbdd0c15992fd75372cbd7fec24244d21b
Parents: 80f4360
Author: Tathagata Das tathagata.das1...@gmail.com
Authored: Mon May 5 15:28:19 2014 -0700
Committer: Tathagata Das tathagata.das1...@gmail.com
Committed: Mon May 5 15:28:54 2014 -0700

--
 docs/configuration.md   |   6 +-
 docs/index.md   |   6 +-
 docs/streaming-custom-receivers.md  | 273 +--
 docs/streaming-programming-guide.md | 200 +++---
 .../streaming/examples/JavaCustomReceiver.java  |   3 +-
 .../streaming/examples/JavaFlumeEventCount.java |   2 +-
 .../streaming/examples/JavaKafkaWordCount.java  |   4 +-
 .../examples/JavaNetworkWordCount.java  |   5 +-
 .../spark/streaming/receiver/Receiver.scala |   6 +-
 .../spark/streaming/StreamingContextSuite.scala |   1 +
 10 files changed, 360 insertions(+), 146 deletions(-)
--


http://git-wip-us.apache.org/repos/asf/spark/blob/1fac4ecb/docs/configuration.md
--
diff --git a/docs/configuration.md b/docs/configuration.md
index 81ad895..d6f316b 100644
--- a/docs/configuration.md
+++ b/docs/configuration.md
@@ -462,7 +462,7 @@ Apart from these, the following properties are also 
available, and may be useful
   td(infinite)/td
   td
 Duration (seconds) of how long Spark will remember any metadata (stages 
generated, tasks generated, etc.).
-Periodic cleanups will ensure that metadata older than this duration will 
be forgetten. This is
+Periodic cleanups will ensure that metadata older than this duration will 
be forgotten. This is
 useful for running Spark for many hours / days (for example, running 24/7 
in case of Spark Streaming
 applications). Note that any RDD that persists in memory for more than 
this duration will be cleared as well.
   /td
@@ -471,8 +471,8 @@ Apart from these, the following properties are also 
available, and may be useful
   tdspark.streaming.blockInterval/td
   td200/td
   td
-Duration (milliseconds) of how long to batch new objects coming from 
network receivers used
-in Spark Streaming.
+Interval (milliseconds) at which data received by Spark Streaming 
receivers is coalesced
+into blocks of data before storing them in Spark.
   /td
 /tr
 tr

http://git-wip-us.apache.org/repos/asf/spark/blob/1fac4ecb/docs/index.md
--
diff --git a/docs/index.md b/docs/index.md
index 6fc9a4f..2daa208 100644
--- a/docs/index.md
+++ b/docs/index.md
@@ -112,10 +112,10 @@ Note that on Windows, you need to set the environment 
variables on separate line
 * [Shark](http://shark.cs.berkeley.edu): Apache Hive over Spark
 * [Mailing Lists](http://spark.apache.org/mailing-lists.html): ask questions 
about Spark here
 * [AMP Camps](http://ampcamp.berkeley.edu/): a series of training camps at UC 
Berkeley that featured talks and
-  exercises about Spark, Shark, Mesos, and more. 
[Videos](http://ampcamp.berkeley.edu/agenda-2012),
-  [slides](http://ampcamp.berkeley.edu/agenda

git commit: Fixed streaming examples docs to use run-example instead of spark-submit

2014-05-14 Thread tdas
Repository: spark
Updated Branches:
  refs/heads/branch-1.0 69ec3149f - c7571d8c6


Fixed streaming examples docs to use run-example instead of spark-submit

Pretty self-explanatory

Author: Tathagata Das tathagata.das1...@gmail.com

Closes #722 from tdas/example-fix and squashes the following commits:

7839979 [Tathagata Das] Minor changes.
0673441 [Tathagata Das] Fixed java docs of java streaming example
e687123 [Tathagata Das] Fixed scala style errors.
9b8d112 [Tathagata Das] Fixed streaming examples docs to use run-example 
instead of spark-submit.


Project: http://git-wip-us.apache.org/repos/asf/spark/repo
Commit: http://git-wip-us.apache.org/repos/asf/spark/commit/c7571d8c
Tree: http://git-wip-us.apache.org/repos/asf/spark/tree/c7571d8c
Diff: http://git-wip-us.apache.org/repos/asf/spark/diff/c7571d8c

Branch: refs/heads/branch-1.0
Commit: c7571d8c6ba058b67cca2b910fd0efacc06642cd
Parents: 69ec314
Author: Tathagata Das tathagata.das1...@gmail.com
Authored: Wed May 14 04:17:32 2014 -0700
Committer: Tathagata Das tathagata.das1...@gmail.com
Committed: Wed May 14 04:24:48 2014 -0700

--
 .../examples/streaming/JavaCustomReceiver.java  | 13 ++---
 .../examples/streaming/JavaFlumeEventCount.java |  6 +-
 .../examples/streaming/JavaKafkaWordCount.java  |  6 +-
 .../streaming/JavaNetworkWordCount.java | 13 +++--
 .../examples/streaming/ActorWordCount.scala |  6 +-
 .../examples/streaming/CustomReceiver.scala | 19 +++
 .../examples/streaming/FlumeEventCount.scala|  9 ++-
 .../examples/streaming/HdfsWordCount.scala  |  5 +-
 .../examples/streaming/KafkaWordCount.scala |  6 +-
 .../examples/streaming/MQTTWordCount.scala  | 10 ++--
 .../examples/streaming/NetworkWordCount.scala   | 14 +++--
 .../streaming/RecoverableNetworkWordCount.scala |  7 +--
 .../streaming/StatefulNetworkWordCount.scala|  6 +-
 .../examples/streaming/TwitterPopularTags.scala | 22 +++-
 .../examples/streaming/ZeroMQWordCount.scala|  8 +--
 .../clickstream/PageViewGenerator.scala | 10 ++--
 .../streaming/clickstream/PageViewStream.scala  |  7 ++-
 .../streaming/twitter/TwitterInputDStream.scala | 58 
 18 files changed, 130 insertions(+), 95 deletions(-)
--


http://git-wip-us.apache.org/repos/asf/spark/blob/c7571d8c/examples/src/main/java/org/apache/spark/examples/streaming/JavaCustomReceiver.java
--
diff --git 
a/examples/src/main/java/org/apache/spark/examples/streaming/JavaCustomReceiver.java
 
b/examples/src/main/java/org/apache/spark/examples/streaming/JavaCustomReceiver.java
index 7f558f3..5622df5 100644
--- 
a/examples/src/main/java/org/apache/spark/examples/streaming/JavaCustomReceiver.java
+++ 
b/examples/src/main/java/org/apache/spark/examples/streaming/JavaCustomReceiver.java
@@ -19,6 +19,7 @@ package org.apache.spark.examples.streaming;
 
 import com.google.common.collect.Lists;
 
+import org.apache.spark.SparkConf;
 import org.apache.spark.api.java.function.FlatMapFunction;
 import org.apache.spark.api.java.function.Function2;
 import org.apache.spark.api.java.function.PairFunction;
@@ -48,25 +49,23 @@ import java.util.regex.Pattern;
  * To run this on your local machine, you need to first run a Netcat server
  *`$ nc -lk `
  * and then run the example
- *`$ ./run org.apache.spark.examples.streaming.JavaCustomReceiver local[2] 
localhost `
+ *`$ bin/run-example 
org.apache.spark.examples.streaming.JavaCustomReceiver localhost `
  */
 
 public class JavaCustomReceiver extends ReceiverString {
   private static final Pattern SPACE = Pattern.compile( );
 
   public static void main(String[] args) {
-if (args.length  3) {
-  System.err.println(Usage: JavaNetworkWordCount master hostname 
port\n +
-  In local mode, master should be 'local[n]' with n  1);
+if (args.length  2) {
+  System.err.println(Usage: JavaNetworkWordCount hostname port);
   System.exit(1);
 }
 
 StreamingExamples.setStreamingLogLevels();
 
 // Create the context with a 1 second batch size
-JavaStreamingContext ssc = new JavaStreamingContext(args[0], 
JavaNetworkWordCount,
-new Duration(1000), System.getenv(SPARK_HOME),
-JavaStreamingContext.jarOfClass(JavaNetworkWordCount.class));
+SparkConf sparkConf = new SparkConf().setAppName(JavaCustomReceiver);
+JavaStreamingContext ssc = new JavaStreamingContext(sparkConf, new 
Duration(1000));
 
 // Create a input stream with the custom receiver on target ip:port and 
count the
 // words in input stream of \n delimited text (eg. generated by 'nc')

http://git-wip-us.apache.org/repos/asf/spark/blob/c7571d8c/examples/src/main/java/org/apache/spark/examples/streaming/JavaFlumeEventCount.java

git commit: Converted bang to ask to avoid scary warning when a block is removed

2014-05-14 Thread tdas
Repository: spark
Updated Branches:
  refs/heads/branch-1.0 1d56cd544 - b8c17e392


Converted bang to ask to avoid scary warning when a block is removed

Removing a block through the blockmanager gave a scary warning messages in the 
driver.
```
2014-05-08 20:16:19,172 WARN BlockManagerMasterActor: Got unknown message: true
2014-05-08 20:16:19,172 WARN BlockManagerMasterActor: Got unknown message: true
2014-05-08 20:16:19,172 WARN BlockManagerMasterActor: Got unknown message: true
```

This is because the 
[BlockManagerSlaveActor](https://github.com/apache/spark/blob/master/core/src/main/scala/org/apache/spark/storage/BlockManagerSlaveActor.scala#L44)
 would send back an acknowledgement (true). But the BlockManagerMasterActor 
would have sent the RemoveBlock message as a send, not as ask(), so would 
reject the receiver true as a unknown message.
@pwendell

Author: Tathagata Das tathagata.das1...@gmail.com

Closes #708 from tdas/bm-fix and squashes the following commits:

ed4ef15 [Tathagata Das] Converted bang to ask to avoid scary warning when a 
block is removed.

(cherry picked from commit 32868f31f88aebd580ab9329dc51a30c26af7a74)
Signed-off-by: Tathagata Das tathagata.das1...@gmail.com


Project: http://git-wip-us.apache.org/repos/asf/spark/repo
Commit: http://git-wip-us.apache.org/repos/asf/spark/commit/b8c17e39
Tree: http://git-wip-us.apache.org/repos/asf/spark/tree/b8c17e39
Diff: http://git-wip-us.apache.org/repos/asf/spark/diff/b8c17e39

Branch: refs/heads/branch-1.0
Commit: b8c17e3928d070d4757d44995516b8872196e5c9
Parents: 1d56cd5
Author: Tathagata Das tathagata.das1...@gmail.com
Authored: Thu May 8 22:34:08 2014 -0700
Committer: Tathagata Das tathagata.das1...@gmail.com
Committed: Thu May 8 22:34:21 2014 -0700

--
 .../scala/org/apache/spark/storage/BlockManagerMasterActor.scala   | 2 +-
 1 file changed, 1 insertion(+), 1 deletion(-)
--


http://git-wip-us.apache.org/repos/asf/spark/blob/b8c17e39/core/src/main/scala/org/apache/spark/storage/BlockManagerMasterActor.scala
--
diff --git 
a/core/src/main/scala/org/apache/spark/storage/BlockManagerMasterActor.scala 
b/core/src/main/scala/org/apache/spark/storage/BlockManagerMasterActor.scala
index 98fa0df..6aed322 100644
--- a/core/src/main/scala/org/apache/spark/storage/BlockManagerMasterActor.scala
+++ b/core/src/main/scala/org/apache/spark/storage/BlockManagerMasterActor.scala
@@ -250,7 +250,7 @@ class BlockManagerMasterActor(val isLocal: Boolean, conf: 
SparkConf, listenerBus
   // Remove the block from the slave's BlockManager.
   // Doesn't actually wait for a confirmation and the message might 
get lost.
   // If message loss becomes frequent, we should add retry logic here.
-  blockManager.get.slaveActor ! RemoveBlock(blockId)
+  blockManager.get.slaveActor.ask(RemoveBlock(blockId))(akkaTimeout)
 }
   }
 }



git commit: Converted bang to ask to avoid scary warning when a block is removed

2014-05-15 Thread tdas
Repository: spark
Updated Branches:
  refs/heads/master 4c60fd1e8 - 32868f31f


Converted bang to ask to avoid scary warning when a block is removed

Removing a block through the blockmanager gave a scary warning messages in the 
driver.
```
2014-05-08 20:16:19,172 WARN BlockManagerMasterActor: Got unknown message: true
2014-05-08 20:16:19,172 WARN BlockManagerMasterActor: Got unknown message: true
2014-05-08 20:16:19,172 WARN BlockManagerMasterActor: Got unknown message: true
```

This is because the 
[BlockManagerSlaveActor](https://github.com/apache/spark/blob/master/core/src/main/scala/org/apache/spark/storage/BlockManagerSlaveActor.scala#L44)
 would send back an acknowledgement (true). But the BlockManagerMasterActor 
would have sent the RemoveBlock message as a send, not as ask(), so would 
reject the receiver true as a unknown message.
@pwendell

Author: Tathagata Das tathagata.das1...@gmail.com

Closes #708 from tdas/bm-fix and squashes the following commits:

ed4ef15 [Tathagata Das] Converted bang to ask to avoid scary warning when a 
block is removed.


Project: http://git-wip-us.apache.org/repos/asf/spark/repo
Commit: http://git-wip-us.apache.org/repos/asf/spark/commit/32868f31
Tree: http://git-wip-us.apache.org/repos/asf/spark/tree/32868f31
Diff: http://git-wip-us.apache.org/repos/asf/spark/diff/32868f31

Branch: refs/heads/master
Commit: 32868f31f88aebd580ab9329dc51a30c26af7a74
Parents: 4c60fd1
Author: Tathagata Das tathagata.das1...@gmail.com
Authored: Thu May 8 22:34:08 2014 -0700
Committer: Tathagata Das tathagata.das1...@gmail.com
Committed: Thu May 8 22:34:08 2014 -0700

--
 .../scala/org/apache/spark/storage/BlockManagerMasterActor.scala   | 2 +-
 1 file changed, 1 insertion(+), 1 deletion(-)
--


http://git-wip-us.apache.org/repos/asf/spark/blob/32868f31/core/src/main/scala/org/apache/spark/storage/BlockManagerMasterActor.scala
--
diff --git 
a/core/src/main/scala/org/apache/spark/storage/BlockManagerMasterActor.scala 
b/core/src/main/scala/org/apache/spark/storage/BlockManagerMasterActor.scala
index 98fa0df..6aed322 100644
--- a/core/src/main/scala/org/apache/spark/storage/BlockManagerMasterActor.scala
+++ b/core/src/main/scala/org/apache/spark/storage/BlockManagerMasterActor.scala
@@ -250,7 +250,7 @@ class BlockManagerMasterActor(val isLocal: Boolean, conf: 
SparkConf, listenerBus
   // Remove the block from the slave's BlockManager.
   // Doesn't actually wait for a confirmation and the message might 
get lost.
   // If message loss becomes frequent, we should add retry logic here.
-  blockManager.get.slaveActor ! RemoveBlock(blockId)
+  blockManager.get.slaveActor.ask(RemoveBlock(blockId))(akkaTimeout)
 }
   }
 }



git commit: [SPARK-1876] Windows fixes to deal with latest distribution layout changes

2014-05-19 Thread tdas
Repository: spark
Updated Branches:
  refs/heads/master df0aa8353 - 7b70a7071


[SPARK-1876] Windows fixes to deal with latest distribution layout changes

- Look for JARs in the right place
- Launch examples the same way as on Unix
- Load datanucleus JARs if they exist
- Don't attempt to parse local paths as URIs in SparkSubmit, since paths with 
C:\ are not valid URIs
- Also fixed POM exclusion rules for datanucleus (it wasn't properly excluding 
it, whereas SBT was)

Author: Matei Zaharia ma...@databricks.com

Closes #819 from mateiz/win-fixes and squashes the following commits:

d558f96 [Matei Zaharia] Fix comment
228577b [Matei Zaharia] Review comments
d3b71c7 [Matei Zaharia] Properly exclude datanucleus files in Maven assembly
144af84 [Matei Zaharia] Update Windows scripts to match latest binary package 
layout


Project: http://git-wip-us.apache.org/repos/asf/spark/repo
Commit: http://git-wip-us.apache.org/repos/asf/spark/commit/7b70a707
Tree: http://git-wip-us.apache.org/repos/asf/spark/tree/7b70a707
Diff: http://git-wip-us.apache.org/repos/asf/spark/diff/7b70a707

Branch: refs/heads/master
Commit: 7b70a7071894dd90ea1d0091542b3e13e7ef8d3a
Parents: df0aa83
Author: Matei Zaharia ma...@databricks.com
Authored: Mon May 19 15:02:35 2014 -0700
Committer: Tathagata Das tathagata.das1...@gmail.com
Committed: Mon May 19 15:02:35 2014 -0700

--
 README.md   |  7 +--
 assembly/pom.xml|  2 +-
 bin/compute-classpath.cmd   | 24 -
 bin/run-example | 23 +
 bin/run-example2.cmd| 51 +++-
 bin/spark-class2.cmd|  2 +
 .../org/apache/spark/deploy/SparkSubmit.scala   |  2 +-
 7 files changed, 81 insertions(+), 30 deletions(-)
--


http://git-wip-us.apache.org/repos/asf/spark/blob/7b70a707/README.md
--
diff --git a/README.md b/README.md
index 9c2e32b..6211a58 100644
--- a/README.md
+++ b/README.md
@@ -9,13 +9,14 @@ You can find the latest Spark documentation, including a 
programming
 guide, on the project webpage at http://spark.apache.org/documentation.html.
 This README file only contains basic setup instructions.
 
-
 ## Building Spark
 
 Spark is built on Scala 2.10. To build Spark and its example programs, run:
 
 ./sbt/sbt assembly
 
+(You do not need to do this if you downloaded a pre-built package.)
+
 ## Interactive Scala Shell
 
 The easiest way to start using Spark is through the Scala shell:
@@ -41,9 +42,9 @@ And run the following command, which should also return 1000:
 Spark also comes with several sample programs in the `examples` directory.
 To run one of them, use `./bin/run-example class [params]`. For example:
 
-./bin/run-example org.apache.spark.examples.SparkLR
+./bin/run-example SparkPi
 
-will run the Logistic Regression example locally.
+will run the Pi example locally.
 
 You can set the MASTER environment variable when running examples to submit
 examples to a cluster. This can be a mesos:// or spark:// URL, 

http://git-wip-us.apache.org/repos/asf/spark/blob/7b70a707/assembly/pom.xml
--
diff --git a/assembly/pom.xml b/assembly/pom.xml
index abd8935..963357b 100644
--- a/assembly/pom.xml
+++ b/assembly/pom.xml
@@ -96,7 +96,7 @@
 filter
   artifact*:*/artifact
   excludes
-excludeorg.datanucleus:*/exclude
+excludeorg/datanucleus/**/exclude
 excludeMETA-INF/*.SF/exclude
 excludeMETA-INF/*.DSA/exclude
 excludeMETA-INF/*.RSA/exclude

http://git-wip-us.apache.org/repos/asf/spark/blob/7b70a707/bin/compute-classpath.cmd
--
diff --git a/bin/compute-classpath.cmd b/bin/compute-classpath.cmd
index 065553e..58710cd 100644
--- a/bin/compute-classpath.cmd
+++ b/bin/compute-classpath.cmd
@@ -20,6 +20,13 @@ rem
 rem This script computes Spark's classpath and prints it to stdout; it's used 
by both the run
 rem script and the ExecutorRunner in standalone cluster mode.
 
+rem If we're called from spark-class2.cmd, it already set 
enabledelayedexpansion and setting
+rem it here would stop us from affecting its copy of the CLASSPATH variable; 
otherwise we
+rem need to set it here because we use !datanucleus_jars! below.
+if %DONT_PRINT_CLASSPATH%==1 goto skip_delayed_expansion
+setlocal enabledelayedexpansion
+:skip_delayed_expansion
+
 set SCALA_VERSION=2.10
 
 rem Figure out where the Spark framework is installed
@@ -31,7 +38,7 @@ if exist %FWDIR%conf\spark-env.cmd call 
%FWDIR%conf\spark-env.cmd
 rem Build up classpath
 set 

git commit: SPARK-1878: Fix the incorrect initialization order

2014-05-19 Thread tdas
Repository: spark
Updated Branches:
  refs/heads/branch-1.0 111c121ae - 901102c1b


SPARK-1878: Fix the incorrect initialization order

JIRA: https://issues.apache.org/jira/browse/SPARK-1878

Author: zsxwing zsxw...@gmail.com

Closes #822 from zsxwing/SPARK-1878 and squashes the following commits:

4a47e27 [zsxwing] SPARK-1878: Fix the incorrect initialization order

(cherry picked from commit 1811ba8ccb580979aa2e12019e6a82805f09ab53)
Signed-off-by: Tathagata Das tathagata.das1...@gmail.com


Project: http://git-wip-us.apache.org/repos/asf/spark/repo
Commit: http://git-wip-us.apache.org/repos/asf/spark/commit/901102c1
Tree: http://git-wip-us.apache.org/repos/asf/spark/tree/901102c1
Diff: http://git-wip-us.apache.org/repos/asf/spark/diff/901102c1

Branch: refs/heads/branch-1.0
Commit: 901102c1ba5f800705819916f2b7a38b6750cffb
Parents: 111c121
Author: zsxwing zsxw...@gmail.com
Authored: Mon May 19 16:41:31 2014 -0700
Committer: Tathagata Das tathagata.das1...@gmail.com
Committed: Mon May 19 16:41:53 2014 -0700

--
 .../apache/spark/streaming/api/java/JavaStreamingContext.scala | 6 +++---
 .../src/test/java/org/apache/spark/streaming/JavaAPISuite.java | 4 
 2 files changed, 7 insertions(+), 3 deletions(-)
--


http://git-wip-us.apache.org/repos/asf/spark/blob/901102c1/streaming/src/main/scala/org/apache/spark/streaming/api/java/JavaStreamingContext.scala
--
diff --git 
a/streaming/src/main/scala/org/apache/spark/streaming/api/java/JavaStreamingContext.scala
 
b/streaming/src/main/scala/org/apache/spark/streaming/api/java/JavaStreamingContext.scala
index 75a3e93..18605ca 100644
--- 
a/streaming/src/main/scala/org/apache/spark/streaming/api/java/JavaStreamingContext.scala
+++ 
b/streaming/src/main/scala/org/apache/spark/streaming/api/java/JavaStreamingContext.scala
@@ -142,12 +142,12 @@ class JavaStreamingContext(val ssc: StreamingContext) {
*/
   def this(path: String, hadoopConf: Configuration) = this(new 
StreamingContext(path, hadoopConf))
 
-  @deprecated(use sparkContext, 0.9.0)
-  val sc: JavaSparkContext = sparkContext
-
   /** The underlying SparkContext */
   val sparkContext = new JavaSparkContext(ssc.sc)
 
+  @deprecated(use sparkContext, 0.9.0)
+  val sc: JavaSparkContext = sparkContext
+
   /**
* Create an input stream from network source hostname:port. Data is 
received using
* a TCP socket and the receive bytes is interpreted as UTF8 encoded \n 
delimited

http://git-wip-us.apache.org/repos/asf/spark/blob/901102c1/streaming/src/test/java/org/apache/spark/streaming/JavaAPISuite.java
--
diff --git 
a/streaming/src/test/java/org/apache/spark/streaming/JavaAPISuite.java 
b/streaming/src/test/java/org/apache/spark/streaming/JavaAPISuite.java
index ce58cb1..4efeb8d 100644
--- a/streaming/src/test/java/org/apache/spark/streaming/JavaAPISuite.java
+++ b/streaming/src/test/java/org/apache/spark/streaming/JavaAPISuite.java
@@ -55,6 +55,10 @@ public class JavaAPISuite extends LocalJavaStreamingContext 
implements Serializa
   equalIterator(a.iterator(), b.iterator());
   }
 
+  @Test
+  public void testInitialization() {
+Assert.assertNotNull(ssc.sc());
+  }
 
   @SuppressWarnings(unchecked)
   @Test



git commit: SPARK-1879. Increase MaxPermSize since some of our builds have many classes

2014-05-19 Thread tdas
Repository: spark
Updated Branches:
  refs/heads/master 1811ba8cc - 5af99d761


SPARK-1879. Increase MaxPermSize since some of our builds have many classes

See https://issues.apache.org/jira/browse/SPARK-1879 -- builds with Hadoop2 and 
Hive ran out of PermGen space in spark-shell, when those things added up with 
the Scala compiler.

Note that users can still override it by setting their own Java options with 
this change. Their options will come later in the command string than the 
-XX:MaxPermSize=128m.

Author: Matei Zaharia ma...@databricks.com

Closes #823 from mateiz/spark-1879 and squashes the following commits:

6bc0ee8 [Matei Zaharia] Increase MaxPermSize to 128m since some of our builds 
have lots of classes


Project: http://git-wip-us.apache.org/repos/asf/spark/repo
Commit: http://git-wip-us.apache.org/repos/asf/spark/commit/5af99d76
Tree: http://git-wip-us.apache.org/repos/asf/spark/tree/5af99d76
Diff: http://git-wip-us.apache.org/repos/asf/spark/diff/5af99d76

Branch: refs/heads/master
Commit: 5af99d7617ba3b9fbfdb345ef9571b7dd41f45a1
Parents: 1811ba8
Author: Matei Zaharia ma...@databricks.com
Authored: Mon May 19 18:42:28 2014 -0700
Committer: Tathagata Das tathagata.das1...@gmail.com
Committed: Mon May 19 18:42:28 2014 -0700

--
 bin/spark-class  | 4 ++--
 bin/spark-class2.cmd | 4 ++--
 .../main/scala/org/apache/spark/deploy/worker/CommandUtils.scala | 4 +++-
 3 files changed, 7 insertions(+), 5 deletions(-)
--


http://git-wip-us.apache.org/repos/asf/spark/blob/5af99d76/bin/spark-class
--
diff --git a/bin/spark-class b/bin/spark-class
index 6480ccb..2e57295 100755
--- a/bin/spark-class
+++ b/bin/spark-class
@@ -99,14 +99,14 @@ else
 fi
 
 # Set JAVA_OPTS to be able to load native libraries and to set heap size
-JAVA_OPTS=$OUR_JAVA_OPTS
+JAVA_OPTS=-XX:MaxPermSize=128m $OUR_JAVA_OPTS
 JAVA_OPTS=$JAVA_OPTS -Xms$OUR_JAVA_MEM -Xmx$OUR_JAVA_MEM
 # Load extra JAVA_OPTS from conf/java-opts, if it exists
 if [ -e $FWDIR/conf/java-opts ] ; then
   JAVA_OPTS=$JAVA_OPTS `cat $FWDIR/conf/java-opts`
 fi
 export JAVA_OPTS
-# Attention: when changing the way the JAVA_OPTS are assembled, the change 
must be reflected in ExecutorRunner.scala!
+# Attention: when changing the way the JAVA_OPTS are assembled, the change 
must be reflected in CommandUtils.scala!
 
 if [ ! -f $FWDIR/RELEASE ]; then
   # Exit if the user hasn't compiled Spark

http://git-wip-us.apache.org/repos/asf/spark/blob/5af99d76/bin/spark-class2.cmd
--
diff --git a/bin/spark-class2.cmd b/bin/spark-class2.cmd
index 266edd9..e420eb4 100755
--- a/bin/spark-class2.cmd
+++ b/bin/spark-class2.cmd
@@ -77,8 +77,8 @@ rem All drivers use SPARK_JAVA_OPTS + SPARK_DRIVER_MEMORY. 
The repl also uses SP
 )
 
 rem Set JAVA_OPTS to be able to load native libraries and to set heap size
-set JAVA_OPTS=%OUR_JAVA_OPTS% -Djava.library.path=%SPARK_LIBRARY_PATH% 
-Xms%OUR_JAVA_MEM% -Xmx%OUR_JAVA_MEM%
-rem Attention: when changing the way the JAVA_OPTS are assembled, the change 
must be reflected in ExecutorRunner.scala!
+set JAVA_OPTS=-XX:MaxPermSize=128m %OUR_JAVA_OPTS% 
-Djava.library.path=%SPARK_LIBRARY_PATH% -Xms%OUR_JAVA_MEM% -Xmx%OUR_JAVA_MEM%
+rem Attention: when changing the way the JAVA_OPTS are assembled, the change 
must be reflected in CommandUtils.scala!
 
 rem Test whether the user has built Spark
 if exist %FWDIR%RELEASE goto skip_build_test

http://git-wip-us.apache.org/repos/asf/spark/blob/5af99d76/core/src/main/scala/org/apache/spark/deploy/worker/CommandUtils.scala
--
diff --git 
a/core/src/main/scala/org/apache/spark/deploy/worker/CommandUtils.scala 
b/core/src/main/scala/org/apache/spark/deploy/worker/CommandUtils.scala
index c7f0f24..4af5bc3 100644
--- a/core/src/main/scala/org/apache/spark/deploy/worker/CommandUtils.scala
+++ b/core/src/main/scala/org/apache/spark/deploy/worker/CommandUtils.scala
@@ -65,6 +65,8 @@ object CommandUtils extends Logging {
  Seq()
   }
 
+val permGenOpt = Seq(-XX:MaxPermSize=128m)
+
 // Figure out our classpath with the external compute-classpath script
 val ext = if (System.getProperty(os.name).startsWith(Windows)) .cmd 
else .sh
 val classPath = Utils.executeAndGetOutput(
@@ -73,7 +75,7 @@ object CommandUtils extends Logging {
 val userClassPath = command.classPathEntries ++ Seq(classPath)
 
 Seq(-cp, 
userClassPath.filterNot(_.isEmpty).mkString(File.pathSeparator)) ++
-  libraryOpts ++ extraOpts ++ workerLocalOpts ++ memoryOpts
+  permGenOpt ++ libraryOpts ++ extraOpts ++ workerLocalOpts ++ memoryOpts
   }
 
   /** Spawn a thread that will 

git commit: SPARK-1689: Spark application should die when removed by Master

2014-05-19 Thread tdas
Repository: spark
Updated Branches:
  refs/heads/master 6a2c5c610 - b0ce22e07


SPARK-1689: Spark application should die when removed by Master

scheduler.error() will mask the error if there are active tasks. Being removed 
is a cataclysmic event for Spark applications, and should probably be treated 
as such.

Author: Aaron Davidson aa...@databricks.com

Closes #832 from aarondav/i-love-u and squashes the following commits:

9f1200f [Aaron Davidson] SPARK-1689: Spark application should die when removed 
by Master


Project: http://git-wip-us.apache.org/repos/asf/spark/repo
Commit: http://git-wip-us.apache.org/repos/asf/spark/commit/b0ce22e0
Tree: http://git-wip-us.apache.org/repos/asf/spark/tree/b0ce22e0
Diff: http://git-wip-us.apache.org/repos/asf/spark/diff/b0ce22e0

Branch: refs/heads/master
Commit: b0ce22e071da4cc62ec5e29abf7b1299b8e4a6b0
Parents: 6a2c5c6
Author: Aaron Davidson aa...@databricks.com
Authored: Mon May 19 20:55:26 2014 -0700
Committer: Tathagata Das tathagata.das1...@gmail.com
Committed: Mon May 19 20:55:26 2014 -0700

--
 .../spark/scheduler/cluster/SparkDeploySchedulerBackend.scala  | 2 ++
 1 file changed, 2 insertions(+)
--


http://git-wip-us.apache.org/repos/asf/spark/blob/b0ce22e0/core/src/main/scala/org/apache/spark/scheduler/cluster/SparkDeploySchedulerBackend.scala
--
diff --git 
a/core/src/main/scala/org/apache/spark/scheduler/cluster/SparkDeploySchedulerBackend.scala
 
b/core/src/main/scala/org/apache/spark/scheduler/cluster/SparkDeploySchedulerBackend.scala
index 9768670..9c07b3f 100644
--- 
a/core/src/main/scala/org/apache/spark/scheduler/cluster/SparkDeploySchedulerBackend.scala
+++ 
b/core/src/main/scala/org/apache/spark/scheduler/cluster/SparkDeploySchedulerBackend.scala
@@ -88,6 +88,8 @@ private[spark] class SparkDeploySchedulerBackend(
 if (!stopping) {
   logError(Application has been killed. Reason:  + reason)
   scheduler.error(reason)
+  // Ensure the application terminates, as we can no longer run jobs.
+  sc.stop()
 }
   }
 



git commit: SPARK-1689: Spark application should die when removed by Master

2014-05-19 Thread tdas
Repository: spark
Updated Branches:
  refs/heads/branch-1.0 875c54fb3 - 78b6e6f1e


SPARK-1689: Spark application should die when removed by Master

scheduler.error() will mask the error if there are active tasks. Being removed 
is a cataclysmic event for Spark applications, and should probably be treated 
as such.

Author: Aaron Davidson aa...@databricks.com

Closes #832 from aarondav/i-love-u and squashes the following commits:

9f1200f [Aaron Davidson] SPARK-1689: Spark application should die when removed 
by Master

(cherry picked from commit b0ce22e071da4cc62ec5e29abf7b1299b8e4a6b0)
Signed-off-by: Tathagata Das tathagata.das1...@gmail.com


Project: http://git-wip-us.apache.org/repos/asf/spark/repo
Commit: http://git-wip-us.apache.org/repos/asf/spark/commit/78b6e6f1
Tree: http://git-wip-us.apache.org/repos/asf/spark/tree/78b6e6f1
Diff: http://git-wip-us.apache.org/repos/asf/spark/diff/78b6e6f1

Branch: refs/heads/branch-1.0
Commit: 78b6e6f1e8ee6a27ef4eed93aac6eba716b5ffce
Parents: 875c54f
Author: Aaron Davidson aa...@databricks.com
Authored: Mon May 19 20:55:26 2014 -0700
Committer: Tathagata Das tathagata.das1...@gmail.com
Committed: Mon May 19 21:01:18 2014 -0700

--
 .../spark/scheduler/cluster/SparkDeploySchedulerBackend.scala  | 2 ++
 1 file changed, 2 insertions(+)
--


http://git-wip-us.apache.org/repos/asf/spark/blob/78b6e6f1/core/src/main/scala/org/apache/spark/scheduler/cluster/SparkDeploySchedulerBackend.scala
--
diff --git 
a/core/src/main/scala/org/apache/spark/scheduler/cluster/SparkDeploySchedulerBackend.scala
 
b/core/src/main/scala/org/apache/spark/scheduler/cluster/SparkDeploySchedulerBackend.scala
index 9768670..9c07b3f 100644
--- 
a/core/src/main/scala/org/apache/spark/scheduler/cluster/SparkDeploySchedulerBackend.scala
+++ 
b/core/src/main/scala/org/apache/spark/scheduler/cluster/SparkDeploySchedulerBackend.scala
@@ -88,6 +88,8 @@ private[spark] class SparkDeploySchedulerBackend(
 if (!stopping) {
   logError(Application has been killed. Reason:  + reason)
   scheduler.error(reason)
+  // Ensure the application terminates, as we can no longer run jobs.
+  sc.stop()
 }
   }
 



[4/4] git commit: [SPARK-1874][MLLIB] Clean up MLlib sample data

2014-05-19 Thread tdas
[SPARK-1874][MLLIB] Clean up MLlib sample data

1. Added synthetic datasets for `MovieLensALS`, `LinearRegression`, 
`BinaryClassification`.
2. Embedded instructions in the help message of those example apps.

Per discussion with Matei on the JIRA page, new example data is under 
`data/mllib`.

Author: Xiangrui Meng m...@databricks.com

Closes #833 from mengxr/mllib-sample-data and squashes the following commits:

59f0a18 [Xiangrui Meng] add sample binary classification data
3c2f92f [Xiangrui Meng] add linear regression data
050f1ca [Xiangrui Meng] add a sample dataset for MovieLensALS example


Project: http://git-wip-us.apache.org/repos/asf/spark/repo
Commit: http://git-wip-us.apache.org/repos/asf/spark/commit/bcb9dce6
Tree: http://git-wip-us.apache.org/repos/asf/spark/tree/bcb9dce6
Diff: http://git-wip-us.apache.org/repos/asf/spark/diff/bcb9dce6

Branch: refs/heads/master
Commit: bcb9dce6f444a977c714117811bce0c54b417650
Parents: b0ce22e
Author: Xiangrui Meng m...@databricks.com
Authored: Mon May 19 21:29:33 2014 -0700
Committer: Tathagata Das tathagata.das1...@gmail.com
Committed: Mon May 19 21:29:33 2014 -0700

--
 .../mllib/sample_binary_classification_data.txt |  100 ++
 data/mllib/sample_linear_regression_data.txt|  501 ++
 data/mllib/sample_movielens_data.txt| 1501 ++
 .../examples/mllib/BinaryClassification.scala   |   12 +-
 .../spark/examples/mllib/LinearRegression.scala |   11 +-
 .../spark/examples/mllib/MovieLensALS.scala |   15 +
 6 files changed, 2138 insertions(+), 2 deletions(-)
--




[1/4] [SPARK-1874][MLLIB] Clean up MLlib sample data

2014-05-19 Thread tdas
Repository: spark
Updated Branches:
  refs/heads/branch-1.0 78b6e6f1e - 1c6c8b5bd


http://git-wip-us.apache.org/repos/asf/spark/blob/1c6c8b5b/data/mllib/sample_movielens_data.txt
--
diff --git a/data/mllib/sample_movielens_data.txt 
b/data/mllib/sample_movielens_data.txt
new file mode 100644
index 000..f0eee19
--- /dev/null
+++ b/data/mllib/sample_movielens_data.txt
@@ -0,0 +1,1501 @@
+0::2::3
+0::3::1
+0::5::2
+0::9::4
+0::11::1
+0::12::2
+0::15::1
+0::17::1
+0::19::1
+0::21::1
+0::23::1
+0::26::3
+0::27::1
+0::28::1
+0::29::1
+0::30::1
+0::31::1
+0::34::1
+0::37::1
+0::41::2
+0::44::1
+0::45::2
+0::46::1
+0::47::1
+0::48::1
+0::50::1
+0::51::1
+0::54::1
+0::55::1
+0::59::2
+0::61::2
+0::64::1
+0::67::1
+0::68::1
+0::69::1
+0::71::1
+0::72::1
+0::77::2
+0::79::1
+0::83::1
+0::87::1
+0::89::2
+0::91::3
+0::92::4
+0::94::1
+0::95::2
+0::96::1
+0::98::1
+0::99::1
+1::2::2
+1::3::1
+1::4::2
+1::6::1
+1::9::3
+1::12::1
+1::13::1
+1::14::1
+1::16::1
+1::19::1
+1::21::3
+1::27::1
+1::28::3
+1::33::1
+1::36::2
+1::37::1
+1::40::1
+1::41::2
+1::43::1
+1::44::1
+1::47::1
+1::50::1
+1::54::1
+1::56::2
+1::57::1
+1::58::1
+1::60::1
+1::62::4
+1::63::1
+1::67::1
+1::68::4
+1::70::2
+1::72::1
+1::73::1
+1::74::2
+1::76::1
+1::77::3
+1::78::1
+1::81::1
+1::82::1
+1::85::3
+1::86::2
+1::88::2
+1::91::1
+1::92::2
+1::93::1
+1::94::2
+1::96::1
+1::97::1
+2::4::3
+2::6::1
+2::8::5
+2::9::1
+2::10::1
+2::12::3
+2::13::1
+2::15::2
+2::18::2
+2::19::4
+2::22::1
+2::26::1
+2::28::1
+2::34::4
+2::35::1
+2::37::5
+2::38::1
+2::39::5
+2::40::4
+2::47::1
+2::50::1
+2::52::2
+2::54::1
+2::55::1
+2::57::2
+2::58::2
+2::59::1
+2::61::1
+2::62::1
+2::64::1
+2::65::1
+2::66::3
+2::68::1
+2::71::3
+2::76::1
+2::77::1
+2::78::1
+2::80::1
+2::83::5
+2::85::1
+2::87::2
+2::88::1
+2::89::4
+2::90::1
+2::92::4
+2::93::5
+3::0::1
+3::1::1
+3::2::1
+3::7::3
+3::8::3
+3::9::1
+3::14::1
+3::15::1
+3::16::1
+3::18::4
+3::19::1
+3::24::3
+3::26::1
+3::29::3
+3::33::1
+3::34::3
+3::35::1
+3::36::3
+3::37::1
+3::38::2
+3::43::1
+3::44::1
+3::46::1
+3::47::1
+3::51::5
+3::52::3
+3::56::1
+3::58::1
+3::60::3
+3::62::1
+3::65::2
+3::66::1
+3::67::1
+3::68::2
+3::70::1
+3::72::2
+3::76::3
+3::79::3
+3::80::4
+3::81::1
+3::83::1
+3::84::1
+3::86::1
+3::87::2
+3::88::4
+3::89::1
+3::91::1
+3::94::3
+4::1::1
+4::6::1
+4::8::1
+4::9::1
+4::10::1
+4::11::1
+4::12::1
+4::13::1
+4::14::2
+4::15::1
+4::17::1
+4::20::1
+4::22::1
+4::23::1
+4::24::1
+4::29::4
+4::30::1
+4::31::1
+4::34::1
+4::35::1
+4::36::1
+4::39::2
+4::40::3
+4::41::4
+4::43::2
+4::44::1
+4::45::1
+4::46::1
+4::47::1
+4::49::2
+4::50::1
+4::51::1
+4::52::4
+4::54::1
+4::55::1
+4::60::3
+4::61::1
+4::62::4
+4::63::3
+4::65::1
+4::67::2
+4::69::1
+4::70::4
+4::71::1
+4::73::1
+4::78::1
+4::84::1
+4::85::1
+4::87::3
+4::88::3
+4::89::2
+4::96::1
+4::97::1
+4::98::1
+4::99::1
+5::0::1
+5::1::1
+5::4::1
+5::5::1
+5::8::1
+5::9::3
+5::10::2
+5::13::3
+5::15::1
+5::19::1
+5::20::3
+5::21::2
+5::23::3
+5::27::1
+5::28::1
+5::29::1
+5::31::1
+5::36::3
+5::38::2
+5::39::1
+5::42::1
+5::48::3
+5::49::4
+5::50::3
+5::51::1
+5::52::1
+5::54::1
+5::55::5
+5::56::3
+5::58::1
+5::60::1
+5::61::1
+5::64::3
+5::65::2
+5::68::4
+5::70::1
+5::71::1
+5::72::1
+5::74::1
+5::79::1
+5::81::2
+5::84::1
+5::85::1
+5::86::1
+5::88::1
+5::90::4
+5::91::2
+5::95::2
+5::99::1
+6::0::1
+6::1::1
+6::2::3
+6::5::1
+6::6::1
+6::9::1
+6::10::1
+6::15::2
+6::16::2
+6::17::1
+6::18::1
+6::20::1
+6::21::1
+6::22::1
+6::24::1
+6::25::5
+6::26::1
+6::28::1
+6::30::1
+6::33::1
+6::38::1
+6::39::1
+6::43::4
+6::44::1
+6::45::1
+6::48::1
+6::49::1
+6::50::1
+6::53::1
+6::54::1
+6::55::1
+6::56::1
+6::58::4
+6::59::1
+6::60::1
+6::61::3
+6::63::3
+6::66::1
+6::67::3
+6::68::1
+6::69::1
+6::71::2
+6::73::1
+6::75::1
+6::77::1
+6::79::1
+6::81::1
+6::84::1
+6::85::3
+6::86::1
+6::87::1
+6::88::1
+6::89::1
+6::91::2
+6::94::1
+6::95::2
+6::96::1
+7::1::1
+7::2::2
+7::3::1
+7::4::1
+7::7::1
+7::10::1
+7::11::2
+7::14::2
+7::15::1
+7::16::1
+7::18::1
+7::21::1
+7::22::1
+7::23::1
+7::25::5
+7::26::1
+7::29::4
+7::30::1
+7::31::3
+7::32::1
+7::33::1
+7::35::1
+7::37::2
+7::39::3
+7::40::2
+7::42::2
+7::44::1
+7::45::2
+7::47::4
+7::48::1
+7::49::1
+7::53::1
+7::54::1
+7::55::1
+7::56::1
+7::59::1
+7::61::2
+7::62::3
+7::63::2
+7::66::1
+7::67::3
+7::74::1
+7::75::1
+7::76::3
+7::77::1
+7::81::1
+7::82::1
+7::84::2
+7::85::4
+7::86::1
+7::92::2
+7::96::1
+7::97::1
+7::98::1
+8::0::1
+8::2::4
+8::3::2
+8::4::2
+8::5::1
+8::7::1
+8::9::1
+8::11::1
+8::15::1
+8::18::1
+8::19::1
+8::21::1
+8::29::5
+8::31::3
+8::33::1
+8::35::1
+8::36::1
+8::40::2
+8::44::1
+8::45::1
+8::50::1
+8::51::1
+8::52::5
+8::53::5
+8::54::1
+8::55::1
+8::56::1
+8::58::4
+8::60::3
+8::62::4
+8::64::1
+8::67::3
+8::69::1
+8::71::1
+8::72::3
+8::77::3
+8::78::1
+8::79::1
+8::83::1
+8::85::5
+8::86::1
+8::88::1
+8::90::1
+8::92::2
+8::95::4
+8::96::3
+8::97::1
+8::98::1
+8::99::1
+9::2::3
+9::3::1

git commit: [Spark 1877] ClassNotFoundException when loading RDD with serialized objects

2014-05-19 Thread tdas
Repository: spark
Updated Branches:
  refs/heads/branch-1.0 1c6c8b5bd - 6cbe2a37c


[Spark 1877] ClassNotFoundException when loading RDD with serialized objects

Updated version of #821

Author: Tathagata Das tathagata.das1...@gmail.com
Author: Ghidireac bogd...@u448a5b0a73d45358d94a.ant.amazon.com

Closes #835 from tdas/SPARK-1877 and squashes the following commits:

f346f71 [Tathagata Das] Addressed Patrick's comments.
fee0c5d [Ghidireac] SPARK-1877: ClassNotFoundException when loading RDD with 
serialized objects

(cherry picked from commit 52eb54d02403a3c37d84b9da7cc1cdb261048cf8)
Signed-off-by: Tathagata Das tathagata.das1...@gmail.com


Project: http://git-wip-us.apache.org/repos/asf/spark/repo
Commit: http://git-wip-us.apache.org/repos/asf/spark/commit/6cbe2a37
Tree: http://git-wip-us.apache.org/repos/asf/spark/tree/6cbe2a37
Diff: http://git-wip-us.apache.org/repos/asf/spark/diff/6cbe2a37

Branch: refs/heads/branch-1.0
Commit: 6cbe2a37ccb14f65b6d6b813a585adbbc43684c4
Parents: 1c6c8b5
Author: Tathagata Das tathagata.das1...@gmail.com
Authored: Mon May 19 22:36:24 2014 -0700
Committer: Tathagata Das tathagata.das1...@gmail.com
Committed: Mon May 19 22:36:37 2014 -0700

--
 core/src/main/scala/org/apache/spark/SparkContext.scala | 2 +-
 1 file changed, 1 insertion(+), 1 deletion(-)
--


http://git-wip-us.apache.org/repos/asf/spark/blob/6cbe2a37/core/src/main/scala/org/apache/spark/SparkContext.scala
--
diff --git a/core/src/main/scala/org/apache/spark/SparkContext.scala 
b/core/src/main/scala/org/apache/spark/SparkContext.scala
index 634c10c..49737fa 100644
--- a/core/src/main/scala/org/apache/spark/SparkContext.scala
+++ b/core/src/main/scala/org/apache/spark/SparkContext.scala
@@ -718,7 +718,7 @@ class SparkContext(config: SparkConf) extends Logging {
   minPartitions: Int = defaultMinPartitions
   ): RDD[T] = {
 sequenceFile(path, classOf[NullWritable], classOf[BytesWritable], 
minPartitions)
-  .flatMap(x = Utils.deserialize[Array[T]](x._2.getBytes))
+  .flatMap(x = Utils.deserialize[Array[T]](x._2.getBytes, 
Utils.getContextOrSparkClassLoader))
   }
 
   protected[spark] def checkpointFile[T: ClassTag](



git commit: [Spark 1877] ClassNotFoundException when loading RDD with serialized objects

2014-05-19 Thread tdas
Repository: spark
Updated Branches:
  refs/heads/master bcb9dce6f - 52eb54d02


[Spark 1877] ClassNotFoundException when loading RDD with serialized objects

Updated version of #821

Author: Tathagata Das tathagata.das1...@gmail.com
Author: Ghidireac bogd...@u448a5b0a73d45358d94a.ant.amazon.com

Closes #835 from tdas/SPARK-1877 and squashes the following commits:

f346f71 [Tathagata Das] Addressed Patrick's comments.
fee0c5d [Ghidireac] SPARK-1877: ClassNotFoundException when loading RDD with 
serialized objects


Project: http://git-wip-us.apache.org/repos/asf/spark/repo
Commit: http://git-wip-us.apache.org/repos/asf/spark/commit/52eb54d0
Tree: http://git-wip-us.apache.org/repos/asf/spark/tree/52eb54d0
Diff: http://git-wip-us.apache.org/repos/asf/spark/diff/52eb54d0

Branch: refs/heads/master
Commit: 52eb54d02403a3c37d84b9da7cc1cdb261048cf8
Parents: bcb9dce
Author: Tathagata Das tathagata.das1...@gmail.com
Authored: Mon May 19 22:36:24 2014 -0700
Committer: Tathagata Das tathagata.das1...@gmail.com
Committed: Mon May 19 22:36:24 2014 -0700

--
 core/src/main/scala/org/apache/spark/SparkContext.scala | 2 +-
 1 file changed, 1 insertion(+), 1 deletion(-)
--


http://git-wip-us.apache.org/repos/asf/spark/blob/52eb54d0/core/src/main/scala/org/apache/spark/SparkContext.scala
--
diff --git a/core/src/main/scala/org/apache/spark/SparkContext.scala 
b/core/src/main/scala/org/apache/spark/SparkContext.scala
index 634c10c..49737fa 100644
--- a/core/src/main/scala/org/apache/spark/SparkContext.scala
+++ b/core/src/main/scala/org/apache/spark/SparkContext.scala
@@ -718,7 +718,7 @@ class SparkContext(config: SparkConf) extends Logging {
   minPartitions: Int = defaultMinPartitions
   ): RDD[T] = {
 sequenceFile(path, classOf[NullWritable], classOf[BytesWritable], 
minPartitions)
-  .flatMap(x = Utils.deserialize[Array[T]](x._2.getBytes))
+  .flatMap(x = Utils.deserialize[Array[T]](x._2.getBytes, 
Utils.getContextOrSparkClassLoader))
   }
 
   protected[spark] def checkpointFile[T: ClassTag](



[2/2] git commit: Revert [maven-release-plugin] prepare release v1.0.0-rc9

2014-05-20 Thread tdas
Revert [maven-release-plugin] prepare release v1.0.0-rc9

This reverts commit 920f947eb5a22a679c0c3186cf69ee75f6041c75.


Project: http://git-wip-us.apache.org/repos/asf/spark/repo
Commit: http://git-wip-us.apache.org/repos/asf/spark/commit/0d988421
Tree: http://git-wip-us.apache.org/repos/asf/spark/tree/0d988421
Diff: http://git-wip-us.apache.org/repos/asf/spark/diff/0d988421

Branch: refs/heads/branch-1.0
Commit: 0d988421742bf43fbd13531fa7ede8d93e59a19b
Parents: 3f3e988
Author: Tathagata Das tathagata.das1...@gmail.com
Authored: Mon May 19 23:15:20 2014 -0700
Committer: Tathagata Das tathagata.das1...@gmail.com
Committed: Mon May 19 23:15:20 2014 -0700

--
 assembly/pom.xml  | 6 +++---
 bagel/pom.xml | 2 +-
 core/pom.xml  | 2 +-
 examples/pom.xml  | 2 +-
 external/flume/pom.xml| 2 +-
 external/kafka/pom.xml| 2 +-
 external/mqtt/pom.xml | 2 +-
 external/twitter/pom.xml  | 2 +-
 external/zeromq/pom.xml   | 2 +-
 extras/spark-ganglia-lgpl/pom.xml | 2 +-
 graphx/pom.xml| 2 +-
 mllib/pom.xml | 2 +-
 pom.xml   | 4 ++--
 repl/pom.xml  | 2 +-
 sql/catalyst/pom.xml  | 2 +-
 sql/core/pom.xml  | 2 +-
 sql/hive/pom.xml  | 2 +-
 streaming/pom.xml | 2 +-
 tools/pom.xml | 2 +-
 yarn/pom.xml  | 2 +-
 yarn/stable/pom.xml   | 2 +-
 21 files changed, 24 insertions(+), 24 deletions(-)
--


http://git-wip-us.apache.org/repos/asf/spark/blob/0d988421/assembly/pom.xml
--
diff --git a/assembly/pom.xml b/assembly/pom.xml
index bed4261..963357b 100644
--- a/assembly/pom.xml
+++ b/assembly/pom.xml
@@ -21,7 +21,7 @@
   parent
 groupIdorg.apache.spark/groupId
 artifactIdspark-parent/artifactId
-version1.0.0/version
+version1.0.0-SNAPSHOT/version
 relativePath../pom.xml/relativePath
   /parent
 
@@ -122,8 +122,8 @@
 transformer 
implementation=org.apache.maven.plugins.shade.resource.DontIncludeResourceTransformer
   resourcelog4j.properties/resource
 /transformer
-transformer 
implementation=org.apache.maven.plugins.shade.resource.ApacheLicenseResourceTransformer
 /
-transformer 
implementation=org.apache.maven.plugins.shade.resource.ApacheNoticeResourceTransformer
 /
+transformer 
implementation=org.apache.maven.plugins.shade.resource.ApacheLicenseResourceTransformer/
+transformer 
implementation=org.apache.maven.plugins.shade.resource.ApacheNoticeResourceTransformer/
   /transformers
 /configuration
   /execution

http://git-wip-us.apache.org/repos/asf/spark/blob/0d988421/bagel/pom.xml
--
diff --git a/bagel/pom.xml b/bagel/pom.xml
index 08932bb..355f437 100644
--- a/bagel/pom.xml
+++ b/bagel/pom.xml
@@ -21,7 +21,7 @@
   parent
 groupIdorg.apache.spark/groupId
 artifactIdspark-parent/artifactId
-version1.0.0/version
+version1.0.0-SNAPSHOT/version
 relativePath../pom.xml/relativePath
   /parent
 

http://git-wip-us.apache.org/repos/asf/spark/blob/0d988421/core/pom.xml
--
diff --git a/core/pom.xml b/core/pom.xml
index 3e22641..bab50f5 100644
--- a/core/pom.xml
+++ b/core/pom.xml
@@ -21,7 +21,7 @@
   parent
 groupIdorg.apache.spark/groupId
 artifactIdspark-parent/artifactId
-version1.0.0/version
+version1.0.0-SNAPSHOT/version
 relativePath../pom.xml/relativePath
   /parent
 

http://git-wip-us.apache.org/repos/asf/spark/blob/0d988421/examples/pom.xml
--
diff --git a/examples/pom.xml b/examples/pom.xml
index 006757a..874bcd7 100644
--- a/examples/pom.xml
+++ b/examples/pom.xml
@@ -21,7 +21,7 @@
   parent
 groupIdorg.apache.spark/groupId
 artifactIdspark-parent/artifactId
-version1.0.0/version
+version1.0.0-SNAPSHOT/version
 relativePath../pom.xml/relativePath
   /parent
 

http://git-wip-us.apache.org/repos/asf/spark/blob/0d988421/external/flume/pom.xml
--
diff --git a/external/flume/pom.xml b/external/flume/pom.xml
index 3ba984e..6aec215 100644
--- a/external/flume/pom.xml
+++ b/external/flume/pom.xml
@@ -21,7 +21,7 @@
   parent
 groupIdorg.apache.spark/groupId
 artifactIdspark-parent/artifactId
-version1.0.0/version
+version1.0.0-SNAPSHOT/version
 relativePath../../pom.xml/relativePath
   /parent
 


[1/2] git commit: Revert [maven-release-plugin] prepare for next development iteration

2014-05-20 Thread tdas
Repository: spark
Updated Branches:
  refs/heads/branch-1.0 1c00f2a25 - 0d9884217


Revert [maven-release-plugin] prepare for next development iteration

This reverts commit f8e611955096c5c1c7db5764b9d2851b1d295f0d.


Project: http://git-wip-us.apache.org/repos/asf/spark/repo
Commit: http://git-wip-us.apache.org/repos/asf/spark/commit/3f3e988c
Tree: http://git-wip-us.apache.org/repos/asf/spark/tree/3f3e988c
Diff: http://git-wip-us.apache.org/repos/asf/spark/diff/3f3e988c

Branch: refs/heads/branch-1.0
Commit: 3f3e988cab4ac350f79ae3e2aadbfd0b5e6938e9
Parents: 1c00f2a
Author: Tathagata Das tathagata.das1...@gmail.com
Authored: Mon May 19 23:13:45 2014 -0700
Committer: Tathagata Das tathagata.das1...@gmail.com
Committed: Mon May 19 23:13:45 2014 -0700

--
 assembly/pom.xml  | 2 +-
 bagel/pom.xml | 2 +-
 core/pom.xml  | 2 +-
 examples/pom.xml  | 2 +-
 external/flume/pom.xml| 2 +-
 external/kafka/pom.xml| 2 +-
 external/mqtt/pom.xml | 2 +-
 external/twitter/pom.xml  | 2 +-
 external/zeromq/pom.xml   | 2 +-
 extras/spark-ganglia-lgpl/pom.xml | 2 +-
 graphx/pom.xml| 2 +-
 mllib/pom.xml | 2 +-
 pom.xml   | 4 ++--
 repl/pom.xml  | 2 +-
 sql/catalyst/pom.xml  | 2 +-
 sql/core/pom.xml  | 2 +-
 sql/hive/pom.xml  | 2 +-
 streaming/pom.xml | 2 +-
 tools/pom.xml | 2 +-
 yarn/pom.xml  | 2 +-
 yarn/stable/pom.xml   | 2 +-
 21 files changed, 22 insertions(+), 22 deletions(-)
--


http://git-wip-us.apache.org/repos/asf/spark/blob/3f3e988c/assembly/pom.xml
--
diff --git a/assembly/pom.xml b/assembly/pom.xml
index 720ca77..bed4261 100644
--- a/assembly/pom.xml
+++ b/assembly/pom.xml
@@ -21,7 +21,7 @@
   parent
 groupIdorg.apache.spark/groupId
 artifactIdspark-parent/artifactId
-version1.0.1-SNAPSHOT/version
+version1.0.0/version
 relativePath../pom.xml/relativePath
   /parent
 

http://git-wip-us.apache.org/repos/asf/spark/blob/3f3e988c/bagel/pom.xml
--
diff --git a/bagel/pom.xml b/bagel/pom.xml
index 85f6d99..08932bb 100644
--- a/bagel/pom.xml
+++ b/bagel/pom.xml
@@ -21,7 +21,7 @@
   parent
 groupIdorg.apache.spark/groupId
 artifactIdspark-parent/artifactId
-version1.0.1-SNAPSHOT/version
+version1.0.0/version
 relativePath../pom.xml/relativePath
   /parent
 

http://git-wip-us.apache.org/repos/asf/spark/blob/3f3e988c/core/pom.xml
--
diff --git a/core/pom.xml b/core/pom.xml
index 47c2507..3e22641 100644
--- a/core/pom.xml
+++ b/core/pom.xml
@@ -21,7 +21,7 @@
   parent
 groupIdorg.apache.spark/groupId
 artifactIdspark-parent/artifactId
-version1.0.1-SNAPSHOT/version
+version1.0.0/version
 relativePath../pom.xml/relativePath
   /parent
 

http://git-wip-us.apache.org/repos/asf/spark/blob/3f3e988c/examples/pom.xml
--
diff --git a/examples/pom.xml b/examples/pom.xml
index b7cbb1a..006757a 100644
--- a/examples/pom.xml
+++ b/examples/pom.xml
@@ -21,7 +21,7 @@
   parent
 groupIdorg.apache.spark/groupId
 artifactIdspark-parent/artifactId
-version1.0.1-SNAPSHOT/version
+version1.0.0/version
 relativePath../pom.xml/relativePath
   /parent
 

http://git-wip-us.apache.org/repos/asf/spark/blob/3f3e988c/external/flume/pom.xml
--
diff --git a/external/flume/pom.xml b/external/flume/pom.xml
index b8fc07f..3ba984e 100644
--- a/external/flume/pom.xml
+++ b/external/flume/pom.xml
@@ -21,7 +21,7 @@
   parent
 groupIdorg.apache.spark/groupId
 artifactIdspark-parent/artifactId
-version1.0.1-SNAPSHOT/version
+version1.0.0/version
 relativePath../../pom.xml/relativePath
   /parent
 

http://git-wip-us.apache.org/repos/asf/spark/blob/3f3e988c/external/kafka/pom.xml
--
diff --git a/external/kafka/pom.xml b/external/kafka/pom.xml
index 9eeb2e1..cb4dd47 100644
--- a/external/kafka/pom.xml
+++ b/external/kafka/pom.xml
@@ -21,7 +21,7 @@
   parent
 groupIdorg.apache.spark/groupId
 artifactIdspark-parent/artifactId
-version1.0.1-SNAPSHOT/version
+version1.0.0/version
 relativePath../../pom.xml/relativePath
   /parent
 

http://git-wip-us.apache.org/repos/asf/spark/blob/3f3e988c/external/mqtt/pom.xml
--
diff --git 

git commit: [Hotfix] Blacklisted flaky HiveCompatibility test

2014-05-20 Thread tdas
Repository: spark
Updated Branches:
  refs/heads/branch-1.0 0d9884217 - b4d93d38d


[Hotfix] Blacklisted flaky HiveCompatibility test

`lateral_view_outer` query sometimes returns a different set of 10 rows.

Author: Tathagata Das tathagata.das1...@gmail.com

Closes #838 from tdas/hive-test-fix2 and squashes the following commits:

9128a0d [Tathagata Das] Blacklisted flaky HiveCompatibility test.

(cherry picked from commit 7f0cfe47f4709843d70ceccc25dee7551206ce0d)
Signed-off-by: Tathagata Das tathagata.das1...@gmail.com


Project: http://git-wip-us.apache.org/repos/asf/spark/repo
Commit: http://git-wip-us.apache.org/repos/asf/spark/commit/b4d93d38
Tree: http://git-wip-us.apache.org/repos/asf/spark/tree/b4d93d38
Diff: http://git-wip-us.apache.org/repos/asf/spark/diff/b4d93d38

Branch: refs/heads/branch-1.0
Commit: b4d93d38d9da61721e64919f95447fafe87bf4d1
Parents: 0d98842
Author: Tathagata Das tathagata.das1...@gmail.com
Authored: Tue May 20 10:27:12 2014 -0700
Committer: Tathagata Das tathagata.das1...@gmail.com
Committed: Tue May 20 10:27:28 2014 -0700

--
 .../spark/sql/hive/execution/HiveCompatibilitySuite.scala  | 6 --
 1 file changed, 4 insertions(+), 2 deletions(-)
--


http://git-wip-us.apache.org/repos/asf/spark/blob/b4d93d38/sql/hive/src/test/scala/org/apache/spark/sql/hive/execution/HiveCompatibilitySuite.scala
--
diff --git 
a/sql/hive/src/test/scala/org/apache/spark/sql/hive/execution/HiveCompatibilitySuite.scala
 
b/sql/hive/src/test/scala/org/apache/spark/sql/hive/execution/HiveCompatibilitySuite.scala
index d83732b..9031abf 100644
--- 
a/sql/hive/src/test/scala/org/apache/spark/sql/hive/execution/HiveCompatibilitySuite.scala
+++ 
b/sql/hive/src/test/scala/org/apache/spark/sql/hive/execution/HiveCompatibilitySuite.scala
@@ -169,7 +169,10 @@ class HiveCompatibilitySuite extends HiveQueryFileTest 
with BeforeAndAfter {
 archive_corrupt,
 
 // No support for case sensitivity is resolution using hive properties atm.
-case_sensitivity
+case_sensitivity,
+
+// Flaky test, Hive sometimes returns different set of 10 rows.
+lateral_view_outer
   )
 
   /**
@@ -476,7 +479,6 @@ class HiveCompatibilitySuite extends HiveQueryFileTest with 
BeforeAndAfter {
 join_view,
 lateral_view,
 lateral_view_cp,
-lateral_view_outer,
 lateral_view_ppd,
 lineage1,
 literal_double,



git commit: [SPARK-1870] Make spark-submit --jars work in yarn-cluster mode.

2014-05-22 Thread tdas
Repository: spark
Updated Branches:
  refs/heads/branch-1.0 75af8bd33 - 79cd26c52


[SPARK-1870] Make spark-submit --jars work in yarn-cluster mode.

Sent secondary jars to distributed cache of all containers and add the cached 
jars to classpath before executors start. Tested on a YARN cluster (CDH-5.0).

`spark-submit --jars` also works in standalone server and `yarn-client`. Thanks 
for @andrewor14 for testing!

I removed Doesn't work for drivers in standalone mode with cluster deploy 
mode. from `spark-submit`'s help message, though we haven't tested mesos yet.

CC: @dbtsai @sryza

Author: Xiangrui Meng m...@databricks.com

Closes #848 from mengxr/yarn-classpath and squashes the following commits:

23e7df4 [Xiangrui Meng] rename spark.jar to __spark__.jar and app.jar to 
__app__.jar to avoid confliction apped $CWD/ and $CWD/* to the classpath remove 
unused methods
a40f6ed [Xiangrui Meng] standalone - cluster
65e04ad [Xiangrui Meng] update spark-submit help message and add a comment for 
yarn-client
11e5354 [Xiangrui Meng] minor changes
3e7e1c4 [Xiangrui Meng] use sparkConf instead of hadoop conf
dc3c825 [Xiangrui Meng] add secondary jars to classpath in yarn

(cherry picked from commit dba314029b4c9d72d7e48a2093b39edd01931f57)
Signed-off-by: Tathagata Das tathagata.das1...@gmail.com


Project: http://git-wip-us.apache.org/repos/asf/spark/repo
Commit: http://git-wip-us.apache.org/repos/asf/spark/commit/79cd26c5
Tree: http://git-wip-us.apache.org/repos/asf/spark/tree/79cd26c5
Diff: http://git-wip-us.apache.org/repos/asf/spark/diff/79cd26c5

Branch: refs/heads/branch-1.0
Commit: 79cd26c528902369b9a2354ba0a0d033e0cf5950
Parents: 75af8bd
Author: Xiangrui Meng m...@databricks.com
Authored: Thu May 22 01:52:50 2014 -0700
Committer: Tathagata Das tathagata.das1...@gmail.com
Committed: Thu May 22 01:53:45 2014 -0700

--
 .../spark/deploy/SparkSubmitArguments.scala |  3 +-
 .../apache/spark/deploy/yarn/ClientBase.scala   | 69 +---
 .../cluster/YarnClientSchedulerBackend.scala|  2 +-
 3 files changed, 19 insertions(+), 55 deletions(-)
--


http://git-wip-us.apache.org/repos/asf/spark/blob/79cd26c5/core/src/main/scala/org/apache/spark/deploy/SparkSubmitArguments.scala
--
diff --git 
a/core/src/main/scala/org/apache/spark/deploy/SparkSubmitArguments.scala 
b/core/src/main/scala/org/apache/spark/deploy/SparkSubmitArguments.scala
index 264d454..0cc05fb 100644
--- a/core/src/main/scala/org/apache/spark/deploy/SparkSubmitArguments.scala
+++ b/core/src/main/scala/org/apache/spark/deploy/SparkSubmitArguments.scala
@@ -326,8 +326,7 @@ private[spark] class SparkSubmitArguments(args: 
Seq[String]) {
 |  --class CLASS_NAME  Your application's main class (for Java 
/ Scala apps).
 |  --name NAME A name of your application.
 |  --jars JARS Comma-separated list of local jars to 
include on the driver
-|  and executor classpaths. Doesn't work 
for drivers in
-|  standalone mode with cluster deploy 
mode.
+|  and executor classpaths.
 |  --py-files PY_FILES Comma-separated list of .zip or .egg 
files to place on the
 |  PYTHONPATH for Python apps.
 |  --files FILES   Comma-separated list of files to be 
placed in the working

http://git-wip-us.apache.org/repos/asf/spark/blob/79cd26c5/yarn/common/src/main/scala/org/apache/spark/deploy/yarn/ClientBase.scala
--
diff --git 
a/yarn/common/src/main/scala/org/apache/spark/deploy/yarn/ClientBase.scala 
b/yarn/common/src/main/scala/org/apache/spark/deploy/yarn/ClientBase.scala
index 27a518c..aeb3f00 100644
--- a/yarn/common/src/main/scala/org/apache/spark/deploy/yarn/ClientBase.scala
+++ b/yarn/common/src/main/scala/org/apache/spark/deploy/yarn/ClientBase.scala
@@ -44,7 +44,7 @@ import org.apache.spark.{Logging, SparkConf, SparkContext}
  * Client submits an application to the YARN ResourceManager.
  *
  * Depending on the deployment mode this will launch one of two application 
master classes:
- * 1. In standalone mode, it will launch an 
[[org.apache.spark.deploy.yarn.ApplicationMaster]]
+ * 1. In cluster mode, it will launch an 
[[org.apache.spark.deploy.yarn.ApplicationMaster]]
  *  which launches a driver program inside of the cluster.
  * 2. In client mode, it will launch an 
[[org.apache.spark.deploy.yarn.ExecutorLauncher]] to
  *  request executors on behalf of a driver running outside of the cluster.
@@ -220,10 +220,11 @@ trait ClientBase extends Logging {
   }
 }
 
+var cachedSecondaryJarLinks = ListBuffer.empty[String]
 

git commit: [SPARK-1897] Respect spark.jars (and --jars) in spark-shell

2014-05-22 Thread tdas
Repository: spark
Updated Branches:
  refs/heads/branch-1.0 a56621622 - 23cc40e39


[SPARK-1897] Respect spark.jars (and --jars) in spark-shell

Spark shell currently overwrites `spark.jars` with `ADD_JARS`. In all modes 
except yarn-cluster, this means the `--jar` flag passed to `bin/spark-shell` is 
also discarded. However, in the 
[docs](http://people.apache.org/~pwendell/spark-1.0.0-rc7-docs/scala-programming-guide.html#initializing-spark),
 we explicitly tell the users to add the jars this way.

Author: Andrew Or andrewo...@gmail.com

Closes #849 from andrewor14/shell-jars and squashes the following commits:

928a7e6 [Andrew Or] ',' - , (minor)
afc357c [Andrew Or] Handle spark.jars ==  in SparkILoop, not SparkSubmit
c6da113 [Andrew Or] Do not set spark.jars to 
d8549f7 [Andrew Or] Respect spark.jars and --jars in spark-shell

(cherry picked from commit 8edbee7d1b4afc192d97ba192a5526affc464205)
Signed-off-by: Tathagata Das tathagata.das1...@gmail.com


Project: http://git-wip-us.apache.org/repos/asf/spark/repo
Commit: http://git-wip-us.apache.org/repos/asf/spark/commit/23cc40e3
Tree: http://git-wip-us.apache.org/repos/asf/spark/tree/23cc40e3
Diff: http://git-wip-us.apache.org/repos/asf/spark/diff/23cc40e3

Branch: refs/heads/branch-1.0
Commit: 23cc40e39acc598816ac46b36463e13587a0dd60
Parents: a566216
Author: Andrew Or andrewo...@gmail.com
Authored: Thu May 22 20:25:41 2014 -0700
Committer: Tathagata Das tathagata.das1...@gmail.com
Committed: Thu May 22 20:25:53 2014 -0700

--
 repl/src/main/scala/org/apache/spark/repl/SparkILoop.scala | 8 +++-
 1 file changed, 7 insertions(+), 1 deletion(-)
--


http://git-wip-us.apache.org/repos/asf/spark/blob/23cc40e3/repl/src/main/scala/org/apache/spark/repl/SparkILoop.scala
--
diff --git a/repl/src/main/scala/org/apache/spark/repl/SparkILoop.scala 
b/repl/src/main/scala/org/apache/spark/repl/SparkILoop.scala
index 296da74..55684e9 100644
--- a/repl/src/main/scala/org/apache/spark/repl/SparkILoop.scala
+++ b/repl/src/main/scala/org/apache/spark/repl/SparkILoop.scala
@@ -993,7 +993,13 @@ object SparkILoop {
   implicit def loopToInterpreter(repl: SparkILoop): SparkIMain = repl.intp
   private def echo(msg: String) = Console println msg
 
-  def getAddedJars: Array[String] = 
Option(System.getenv(ADD_JARS)).map(_.split(',')).getOrElse(new 
Array[String](0))
+  def getAddedJars: Array[String] = {
+val envJars = sys.env.get(ADD_JARS)
+val propJars = sys.props.get(spark.jars).flatMap { p =
+  if (p == ) None else Some(p)
+}
+propJars.orElse(envJars).map(_.split(,)).getOrElse(Array.empty)
+  }
 
   // Designed primarily for use by test code: take a String with a
   // bunch of code, and prints out a transcript of what it would look



git commit: [SPARK-1896] Respect spark.master (and --master) before MASTER in spark-shell

2014-05-22 Thread tdas
Repository: spark
Updated Branches:
  refs/heads/master 8edbee7d1 - cce77457e


[SPARK-1896] Respect spark.master (and --master) before MASTER in spark-shell

The hierarchy for configuring the Spark master in the shell is as follows:
```
MASTER  --master  spark.master (spark-defaults.conf)
```
This is inconsistent with the way we run normal applications, which is:
```
--master  spark.master (spark-defaults.conf)  MASTER
```

I was trying to run a shell locally on a standalone cluster launched through 
the ec2 scripts, which automatically set `MASTER` in spark-env.sh. It was 
surprising to me that `--master` didn't take effect, considering that this is 
the way we tell users to set their masters 
[here](http://people.apache.org/~pwendell/spark-1.0.0-rc7-docs/scala-programming-guide.html#initializing-spark).

Author: Andrew Or andrewo...@gmail.com

Closes #846 from andrewor14/shell-master and squashes the following commits:

2cb81c9 [Andrew Or] Respect spark.master before MASTER in REPL


Project: http://git-wip-us.apache.org/repos/asf/spark/repo
Commit: http://git-wip-us.apache.org/repos/asf/spark/commit/cce77457
Tree: http://git-wip-us.apache.org/repos/asf/spark/tree/cce77457
Diff: http://git-wip-us.apache.org/repos/asf/spark/diff/cce77457

Branch: refs/heads/master
Commit: cce77457e00aa5f1f4db3d50454cf257efb156ed
Parents: 8edbee7
Author: Andrew Or andrewo...@gmail.com
Authored: Thu May 22 20:32:27 2014 -0700
Committer: Tathagata Das tathagata.das1...@gmail.com
Committed: Thu May 22 20:32:27 2014 -0700

--
 repl/src/main/scala/org/apache/spark/repl/SparkILoop.scala | 5 ++---
 1 file changed, 2 insertions(+), 3 deletions(-)
--


http://git-wip-us.apache.org/repos/asf/spark/blob/cce77457/repl/src/main/scala/org/apache/spark/repl/SparkILoop.scala
--
diff --git a/repl/src/main/scala/org/apache/spark/repl/SparkILoop.scala 
b/repl/src/main/scala/org/apache/spark/repl/SparkILoop.scala
index 55684e9..5f34362 100644
--- a/repl/src/main/scala/org/apache/spark/repl/SparkILoop.scala
+++ b/repl/src/main/scala/org/apache/spark/repl/SparkILoop.scala
@@ -962,11 +962,10 @@ class SparkILoop(in0: Option[BufferedReader], protected 
val out: JPrintWriter,
   private def getMaster(): String = {
 val master = this.master match {
   case Some(m) = m
-  case None = {
+  case None =
 val envMaster = sys.env.get(MASTER)
 val propMaster = sys.props.get(spark.master)
-envMaster.orElse(propMaster).getOrElse(local[*])
-  }
+propMaster.orElse(envMaster).getOrElse(local[*])
 }
 master
   }



git commit: Updated scripts for auditing releases

2014-05-22 Thread tdas
Repository: spark
Updated Branches:
  refs/heads/master cce77457e - b2bdd0e50


Updated scripts for auditing releases

- Added script to automatically generate change list CHANGES.txt
- Added test for verifying linking against maven distributions of `spark-sql` 
and `spark-hive`
- Added SBT projects for testing functionality of `spark-sql` and `spark-hive`
- Fixed issues in existing tests that might have come up because of changes in 
Spark 1.0

Author: Tathagata Das tathagata.das1...@gmail.com

Closes #844 from tdas/update-dev-scripts and squashes the following commits:

25090ba [Tathagata Das] Added missing license
e2e20b3 [Tathagata Das] Updated tests for auditing releases.


Project: http://git-wip-us.apache.org/repos/asf/spark/repo
Commit: http://git-wip-us.apache.org/repos/asf/spark/commit/b2bdd0e5
Tree: http://git-wip-us.apache.org/repos/asf/spark/tree/b2bdd0e5
Diff: http://git-wip-us.apache.org/repos/asf/spark/diff/b2bdd0e5

Branch: refs/heads/master
Commit: b2bdd0e505f1ae3d39c46139f17bd43779ece635
Parents: cce7745
Author: Tathagata Das tathagata.das1...@gmail.com
Authored: Thu May 22 20:48:55 2014 -0700
Committer: Tathagata Das tathagata.das1...@gmail.com
Committed: Thu May 22 20:48:55 2014 -0700

--
 dev/audit-release/audit_release.py  |  11 +-
 .../maven_app_core/src/main/java/SimpleApp.java |   1 +
 .../sbt_app_core/src/main/scala/SparkApp.scala  |   2 +-
 dev/audit-release/sbt_app_hive/build.sbt|  29 +++
 dev/audit-release/sbt_app_hive/data.txt |   9 +
 .../src/main/resources/hive-site.xml| 213 +++
 .../sbt_app_hive/src/main/scala/HiveApp.scala   |  57 +
 dev/audit-release/sbt_app_sql/build.sbt |  29 +++
 .../sbt_app_sql/src/main/scala/SqlApp.scala |  57 +
 .../src/main/scala/StreamingApp.scala   |   1 -
 dev/create-release/generate-changelist.py   | 144 +
 11 files changed, 547 insertions(+), 6 deletions(-)
--


http://git-wip-us.apache.org/repos/asf/spark/blob/b2bdd0e5/dev/audit-release/audit_release.py
--
diff --git a/dev/audit-release/audit_release.py 
b/dev/audit-release/audit_release.py
index 4a816d4..8c7573b 100755
--- a/dev/audit-release/audit_release.py
+++ b/dev/audit-release/audit_release.py
@@ -93,9 +93,12 @@ original_dir = os.getcwd()
 # For each of these modules, we'll test an 'empty' application in sbt and 
 # maven that links against them. This will catch issues with messed up
 # dependencies within those projects.
-modules = [spark-core, spark-bagel, spark-mllib, spark-streaming, 
spark-repl, 
-   spark-graphx, spark-streaming-flume, spark-streaming-kafka, 
-   spark-streaming-mqtt, spark-streaming-twitter, 
spark-streaming-zeromq]
+modules = [
+  spark-core, spark-bagel, spark-mllib, spark-streaming, spark-repl,
+  spark-graphx, spark-streaming-flume, spark-streaming-kafka,
+  spark-streaming-mqtt, spark-streaming-twitter, spark-streaming-zeromq,
+  spark-catalyst, spark-sql, spark-hive
+]
 modules = map(lambda m: %s_%s % (m, SCALA_BINARY_VERSION), modules)
 
 # Check for directories that might interfere with tests
@@ -122,7 +125,7 @@ for module in modules:
 os.chdir(original_dir)
 
 # SBT application tests
-for app in [sbt_app_core, sbt_app_graphx, sbt_app_streaming]:
+for app in [sbt_app_core, sbt_app_graphx, sbt_app_streaming, 
sbt_app_sql, sbt_app_hive]:
   os.chdir(app)
   ret = run_cmd(sbt clean run, exit_on_failure=False)
   test(ret == 0, sbt application (%s) % app)

http://git-wip-us.apache.org/repos/asf/spark/blob/b2bdd0e5/dev/audit-release/maven_app_core/src/main/java/SimpleApp.java
--
diff --git a/dev/audit-release/maven_app_core/src/main/java/SimpleApp.java 
b/dev/audit-release/maven_app_core/src/main/java/SimpleApp.java
index 6b65dda..5217689 100644
--- a/dev/audit-release/maven_app_core/src/main/java/SimpleApp.java
+++ b/dev/audit-release/maven_app_core/src/main/java/SimpleApp.java
@@ -37,5 +37,6 @@ public class SimpleApp {
  System.exit(-1);
}
System.out.println(Test succeeded);
+   sc.stop();
   }
 }

http://git-wip-us.apache.org/repos/asf/spark/blob/b2bdd0e5/dev/audit-release/sbt_app_core/src/main/scala/SparkApp.scala
--
diff --git a/dev/audit-release/sbt_app_core/src/main/scala/SparkApp.scala 
b/dev/audit-release/sbt_app_core/src/main/scala/SparkApp.scala
index a89b0d7..77bbd16 100644
--- a/dev/audit-release/sbt_app_core/src/main/scala/SparkApp.scala
+++ b/dev/audit-release/sbt_app_core/src/main/scala/SparkApp.scala
@@ -19,6 +19,7 @@ package main.scala
 
 import scala.util.Try
 
+import org.apache.spark.SparkConf
 import org.apache.spark.SparkContext
 import org.apache.spark.SparkContext

git commit: Updated scripts for auditing releases

2014-05-22 Thread tdas
Repository: spark
Updated Branches:
  refs/heads/branch-1.0 c3b406518 - 6541ca24b


Updated scripts for auditing releases

- Added script to automatically generate change list CHANGES.txt
- Added test for verifying linking against maven distributions of `spark-sql` 
and `spark-hive`
- Added SBT projects for testing functionality of `spark-sql` and `spark-hive`
- Fixed issues in existing tests that might have come up because of changes in 
Spark 1.0

Author: Tathagata Das tathagata.das1...@gmail.com

Closes #844 from tdas/update-dev-scripts and squashes the following commits:

25090ba [Tathagata Das] Added missing license
e2e20b3 [Tathagata Das] Updated tests for auditing releases.

(cherry picked from commit b2bdd0e505f1ae3d39c46139f17bd43779ece635)
Signed-off-by: Tathagata Das tathagata.das1...@gmail.com


Project: http://git-wip-us.apache.org/repos/asf/spark/repo
Commit: http://git-wip-us.apache.org/repos/asf/spark/commit/6541ca24
Tree: http://git-wip-us.apache.org/repos/asf/spark/tree/6541ca24
Diff: http://git-wip-us.apache.org/repos/asf/spark/diff/6541ca24

Branch: refs/heads/branch-1.0
Commit: 6541ca24b2cf79d6914b9d0bcf7b51bff405369e
Parents: c3b4065
Author: Tathagata Das tathagata.das1...@gmail.com
Authored: Thu May 22 20:48:55 2014 -0700
Committer: Tathagata Das tathagata.das1...@gmail.com
Committed: Thu May 22 20:49:25 2014 -0700

--
 dev/audit-release/audit_release.py  |  11 +-
 .../maven_app_core/src/main/java/SimpleApp.java |   1 +
 .../sbt_app_core/src/main/scala/SparkApp.scala  |   2 +-
 dev/audit-release/sbt_app_hive/build.sbt|  29 +++
 dev/audit-release/sbt_app_hive/data.txt |   9 +
 .../src/main/resources/hive-site.xml| 213 +++
 .../sbt_app_hive/src/main/scala/HiveApp.scala   |  57 +
 dev/audit-release/sbt_app_sql/build.sbt |  29 +++
 .../sbt_app_sql/src/main/scala/SqlApp.scala |  57 +
 .../src/main/scala/StreamingApp.scala   |   1 -
 dev/create-release/generate-changelist.py   | 144 +
 11 files changed, 547 insertions(+), 6 deletions(-)
--


http://git-wip-us.apache.org/repos/asf/spark/blob/6541ca24/dev/audit-release/audit_release.py
--
diff --git a/dev/audit-release/audit_release.py 
b/dev/audit-release/audit_release.py
index 4a816d4..8c7573b 100755
--- a/dev/audit-release/audit_release.py
+++ b/dev/audit-release/audit_release.py
@@ -93,9 +93,12 @@ original_dir = os.getcwd()
 # For each of these modules, we'll test an 'empty' application in sbt and 
 # maven that links against them. This will catch issues with messed up
 # dependencies within those projects.
-modules = [spark-core, spark-bagel, spark-mllib, spark-streaming, 
spark-repl, 
-   spark-graphx, spark-streaming-flume, spark-streaming-kafka, 
-   spark-streaming-mqtt, spark-streaming-twitter, 
spark-streaming-zeromq]
+modules = [
+  spark-core, spark-bagel, spark-mllib, spark-streaming, spark-repl,
+  spark-graphx, spark-streaming-flume, spark-streaming-kafka,
+  spark-streaming-mqtt, spark-streaming-twitter, spark-streaming-zeromq,
+  spark-catalyst, spark-sql, spark-hive
+]
 modules = map(lambda m: %s_%s % (m, SCALA_BINARY_VERSION), modules)
 
 # Check for directories that might interfere with tests
@@ -122,7 +125,7 @@ for module in modules:
 os.chdir(original_dir)
 
 # SBT application tests
-for app in [sbt_app_core, sbt_app_graphx, sbt_app_streaming]:
+for app in [sbt_app_core, sbt_app_graphx, sbt_app_streaming, 
sbt_app_sql, sbt_app_hive]:
   os.chdir(app)
   ret = run_cmd(sbt clean run, exit_on_failure=False)
   test(ret == 0, sbt application (%s) % app)

http://git-wip-us.apache.org/repos/asf/spark/blob/6541ca24/dev/audit-release/maven_app_core/src/main/java/SimpleApp.java
--
diff --git a/dev/audit-release/maven_app_core/src/main/java/SimpleApp.java 
b/dev/audit-release/maven_app_core/src/main/java/SimpleApp.java
index 6b65dda..5217689 100644
--- a/dev/audit-release/maven_app_core/src/main/java/SimpleApp.java
+++ b/dev/audit-release/maven_app_core/src/main/java/SimpleApp.java
@@ -37,5 +37,6 @@ public class SimpleApp {
  System.exit(-1);
}
System.out.println(Test succeeded);
+   sc.stop();
   }
 }

http://git-wip-us.apache.org/repos/asf/spark/blob/6541ca24/dev/audit-release/sbt_app_core/src/main/scala/SparkApp.scala
--
diff --git a/dev/audit-release/sbt_app_core/src/main/scala/SparkApp.scala 
b/dev/audit-release/sbt_app_core/src/main/scala/SparkApp.scala
index a89b0d7..77bbd16 100644
--- a/dev/audit-release/sbt_app_core/src/main/scala/SparkApp.scala
+++ b/dev/audit-release/sbt_app_core/src/main/scala/SparkApp.scala
@@ -19,6 +19,7 @@ package main.scala
 
 import

git commit: SPARK-1911: Emphasize that Spark jars should be built with Java 6.

2014-05-24 Thread tdas
Repository: spark
Updated Branches:
  refs/heads/master 5081a0a9d - 75a032777


SPARK-1911: Emphasize that Spark jars should be built with Java 6.

This commit requires the user to manually say yes when buiding Spark
without Java 6. The prompt can be bypassed with a flag (e.g. if the user
is scripting around make-distribution).

Author: Patrick Wendell pwend...@gmail.com

Closes #859 from pwendell/java6 and squashes the following commits:

4921133 [Patrick Wendell] Adding Pyspark Notice
fee8c9e [Patrick Wendell] SPARK-1911: Emphasize that Spark jars should be built 
with Java 6.


Project: http://git-wip-us.apache.org/repos/asf/spark/repo
Commit: http://git-wip-us.apache.org/repos/asf/spark/commit/75a03277
Tree: http://git-wip-us.apache.org/repos/asf/spark/tree/75a03277
Diff: http://git-wip-us.apache.org/repos/asf/spark/diff/75a03277

Branch: refs/heads/master
Commit: 75a03277704f8618a0f1c41aecfb1ebd24a8ac1a
Parents: 5081a0a
Author: Patrick Wendell pwend...@gmail.com
Authored: Sat May 24 18:27:00 2014 -0700
Committer: Tathagata Das tathagata.das1...@gmail.com
Committed: Sat May 24 18:27:00 2014 -0700

--
 make-distribution.sh | 52 ---
 1 file changed, 31 insertions(+), 21 deletions(-)
--


http://git-wip-us.apache.org/repos/asf/spark/blob/75a03277/make-distribution.sh
--
diff --git a/make-distribution.sh b/make-distribution.sh
index 1c89027..ae52b49 100755
--- a/make-distribution.sh
+++ b/make-distribution.sh
@@ -46,27 +46,6 @@ set -e
 FWDIR=$(cd `dirname $0`; pwd)
 DISTDIR=$FWDIR/dist
 
-if [ -z $JAVA_HOME ]; then
-  echo Error: JAVA_HOME is not set, cannot proceed.
-  exit -1
-fi
-
-JAVA_CMD=$JAVA_HOME/bin/java
-JAVA_VERSION=$($JAVA_CMD -version 21)
-if ! [[ $JAVA_VERSION =~ 1.6 ]]; then
-  echo ***NOTE***: JAVA_HOME is not set to a JDK 6 installation. The 
resulting
-  echo distribution will not support Java 6. See SPARK-1703.
-  echo Output from 'java -version' was:
-  echo $JAVA_VERSION
-fi
-
-VERSION=$(mvn help:evaluate -Dexpression=project.version 2/dev/null | grep -v 
INFO | tail -n 1)
-if [ $? != 0 ]; then
-echo -e You need Maven installed to build Spark.
-echo -e Download Maven from https://maven.apache.org/;
-exit -1;
-fi
-
 # Initialize defaults
 SPARK_HADOOP_VERSION=1.0.4
 SPARK_YARN=false
@@ -88,6 +67,9 @@ while (( $# )); do
 --with-hive)
   SPARK_HIVE=true
   ;;
+--skip-java-test)
+  SKIP_JAVA_TEST=true
+  ;;
 --with-tachyon)
   SPARK_TACHYON=true
   ;;
@@ -102,6 +84,34 @@ while (( $# )); do
   shift
 done
 
+if [ -z $JAVA_HOME ]; then
+  echo Error: JAVA_HOME is not set, cannot proceed.
+  exit -1
+fi
+
+VERSION=$(mvn help:evaluate -Dexpression=project.version 2/dev/null | grep -v 
INFO | tail -n 1)
+if [ $? != 0 ]; then
+echo -e You need Maven installed to build Spark.
+echo -e Download Maven from https://maven.apache.org/;
+exit -1;
+fi
+
+JAVA_CMD=$JAVA_HOME/bin/java
+JAVA_VERSION=$($JAVA_CMD -version 21)
+if [[ ! $JAVA_VERSION =~ 1.6  -z $SKIP_JAVA_TEST ]]; then
+  echo ***NOTE***: JAVA_HOME is not set to a JDK 6 installation. The 
resulting
+  echo distribution may not work well with PySpark and will not 
run
+  echo with Java 6 (See SPARK-1703 and SPARK-1911).
+  echo This test can be disabled by adding --skip-java-test.
+  echo Output from 'java -version' was:
+  echo $JAVA_VERSION
+  read -p Would you like to continue anyways? [y,n]:  -r
+  if [[ ! $REPLY =~ ^[Yy]$ ]]; then
+echo Okay, exiting.
+exit 1
+  fi 
+fi
+
 if [ $NAME == none ]; then
   NAME=$SPARK_HADOOP_VERSION
 fi



git commit: Updated CHANGES.txt

2014-05-24 Thread tdas
Repository: spark
Updated Branches:
  refs/heads/branch-1.0 217bd5627 - 840609272


Updated CHANGES.txt


Project: http://git-wip-us.apache.org/repos/asf/spark/repo
Commit: http://git-wip-us.apache.org/repos/asf/spark/commit/84060927
Tree: http://git-wip-us.apache.org/repos/asf/spark/tree/84060927
Diff: http://git-wip-us.apache.org/repos/asf/spark/diff/84060927

Branch: refs/heads/branch-1.0
Commit: 84060927262159c4c847034a50ff8092f0a72a44
Parents: 217bd56
Author: Tathagata Das tathagata.das1...@gmail.com
Authored: Sun May 25 02:20:13 2014 +
Committer: Tathagata Das tathagata.das1...@gmail.com
Committed: Sun May 25 02:20:13 2014 +

--
 CHANGES.txt | 87 +++-
 1 file changed, 86 insertions(+), 1 deletion(-)
--


http://git-wip-us.apache.org/repos/asf/spark/blob/84060927/CHANGES.txt
--
diff --git a/CHANGES.txt b/CHANGES.txt
index d21f0ac..b097339 100644
--- a/CHANGES.txt
+++ b/CHANGES.txt
@@ -3,6 +3,91 @@ Spark Change Log
 
 Release 1.0.0
 
+  SPARK-1911: Emphasize that Spark jars should be built with Java 6.
+  Patrick Wendell pwend...@gmail.com
+  2014-05-24 18:27:00 -0700
+  Commit: 217bd56, github.com/apache/spark/pull/859
+
+  [SPARK-1900 / 1918] PySpark on YARN is broken
+  Andrew Or andrewo...@gmail.com
+  2014-05-24 18:01:49 -0700
+  Commit: 12f5ecc, github.com/apache/spark/pull/853
+
+  Update LBFGSSuite.scala
+  baishuo(白硕) vc_j...@hotmail.com
+  2014-05-23 13:02:40 -0700
+  Commit: 9be103a, github.com/apache/spark/pull/815
+
+  Updated scripts for auditing releases
+  Tathagata Das tathagata.das1...@gmail.com
+  2014-05-22 20:48:55 -0700
+  Commit: 6541ca2, github.com/apache/spark/pull/844
+
+  [SPARK-1896] Respect spark.master (and --master) before MASTER in spark-shell
+  Andrew Or andrewo...@gmail.com
+  2014-05-22 20:32:27 -0700
+  Commit: c3b4065, github.com/apache/spark/pull/846
+
+  [SPARK-1897] Respect spark.jars (and --jars) in spark-shell
+  Andrew Or andrewo...@gmail.com
+  2014-05-22 20:25:41 -0700
+  Commit: 23cc40e, github.com/apache/spark/pull/849
+
+  Fix UISuite unit test that fails under Jenkins contention
+  Aaron Davidson aa...@databricks.com
+  2014-05-22 15:11:05 -0700
+  Commit: a566216, github.com/apache/spark/pull/857
+
+  [SPARK-1870] Make spark-submit --jars work in yarn-cluster mode.
+  Xiangrui Meng m...@databricks.com
+  2014-05-22 01:52:50 -0700
+  Commit: 79cd26c, github.com/apache/spark/pull/848
+
+  Configuration documentation updates
+  Reynold Xin r...@apache.org
+  2014-05-21 18:49:12 -0700
+  Commit: 75af8bd, github.com/apache/spark/pull/851
+
+  [SPARK-1889] [SQL] Apply splitConjunctivePredicates to join condition while 
finding join ke...
+  Takuya UESHIN ues...@happy-camper.st
+  2014-05-21 15:37:47 -0700
+  Commit: 6e7934e, github.com/apache/spark/pull/836
+
+  [SPARK-1519] Support minPartitions param of wholeTextFiles() in PySpark
+  Kan Zhang kzh...@apache.org
+  2014-05-21 13:26:53 -0700
+  Commit: 30d1df5, github.com/apache/spark/pull/697
+
+  [Typo] Stoped - Stopped
+  Andrew Or andrewo...@gmail.com
+  2014-05-21 11:59:05 -0700
+  Commit: 9b8f772, github.com/apache/spark/pull/847
+
+  [Minor] Move JdbcRDDSuite to the correct package
+  Andrew Or andrewo...@gmail.com
+  2014-05-21 01:25:10 -0700
+  Commit: bc6bbfa, github.com/apache/spark/pull/839
+
+  [Docs] Correct example of creating a new SparkConf
+  Andrew Or andrewo...@gmail.com
+  2014-05-21 01:23:34 -0700
+  Commit: 7295dd9, github.com/apache/spark/pull/842
+
+  [SPARK-1250] Fixed misleading comments in bin/pyspark, bin/spark-class
+  Sumedh Mungee smun...@gmail.com
+  2014-05-21 01:22:25 -0700
+  Commit: 364c14a, github.com/apache/spark/pull/843
+
+  [Hotfix] Blacklisted flaky HiveCompatibility test
+  Tathagata Das tathagata.das1...@gmail.com
+  2014-05-20 10:27:12 -0700
+  Commit: b4d93d3, github.com/apache/spark/pull/838
+
+  Updated CHANGES.txt
+  Tathagata Das tathagata.das1...@gmail.com
+  2014-05-19 23:12:24 -0700
+  Commit: 1c00f2a
+
   [Spark 1877] ClassNotFoundException when loading RDD with serialized objects
   Tathagata Das tathagata.das1...@gmail.com, Ghidireac 
bogd...@u448a5b0a73d45358d94a.ant.amazon.com
   2014-05-19 22:36:24 -0700
@@ -676,7 +761,7 @@ Release 1.0.0
   Proposal: clarify Scala programming guide on caching ...
   Ethan Jewett esjew...@gmail.com
   2014-05-06 20:50:08 -0700
-  Commit: 51e27755, github.com/apache/spark/pull/668
+  Commit: 51e2775, github.com/apache/spark/pull/668
 
   SPARK-1727. Correct small compile errors, typos, and markdown issues in 
(primarly) MLlib docs
   Sean Owen so...@cloudera.com



git commit: [maven-release-plugin] prepare release v1.0.0-rc11

2014-05-25 Thread tdas
Repository: spark
Updated Branches:
  refs/heads/branch-1.0 7273bfc03 - 2f1dc868e


[maven-release-plugin] prepare release v1.0.0-rc11


Project: http://git-wip-us.apache.org/repos/asf/spark/repo
Commit: http://git-wip-us.apache.org/repos/asf/spark/commit/2f1dc868
Tree: http://git-wip-us.apache.org/repos/asf/spark/tree/2f1dc868
Diff: http://git-wip-us.apache.org/repos/asf/spark/diff/2f1dc868

Branch: refs/heads/branch-1.0
Commit: 2f1dc868e5714882cf40d2633fb66772baf34789
Parents: 7273bfc
Author: Tathagata Das tathagata.das1...@gmail.com
Authored: Sun May 25 10:18:41 2014 +
Committer: Tathagata Das tathagata.das1...@gmail.com
Committed: Sun May 25 10:18:41 2014 +

--
 assembly/pom.xml  | 6 +++---
 bagel/pom.xml | 2 +-
 core/pom.xml  | 2 +-
 examples/pom.xml  | 2 +-
 external/flume/pom.xml| 2 +-
 external/kafka/pom.xml| 2 +-
 external/mqtt/pom.xml | 2 +-
 external/twitter/pom.xml  | 2 +-
 external/zeromq/pom.xml   | 2 +-
 extras/spark-ganglia-lgpl/pom.xml | 2 +-
 graphx/pom.xml| 2 +-
 mllib/pom.xml | 2 +-
 pom.xml   | 4 ++--
 repl/pom.xml  | 2 +-
 sql/catalyst/pom.xml  | 2 +-
 sql/core/pom.xml  | 2 +-
 sql/hive/pom.xml  | 2 +-
 streaming/pom.xml | 2 +-
 tools/pom.xml | 2 +-
 yarn/pom.xml  | 2 +-
 yarn/stable/pom.xml   | 2 +-
 21 files changed, 24 insertions(+), 24 deletions(-)
--


http://git-wip-us.apache.org/repos/asf/spark/blob/2f1dc868/assembly/pom.xml
--
diff --git a/assembly/pom.xml b/assembly/pom.xml
index 963357b..bed4261 100644
--- a/assembly/pom.xml
+++ b/assembly/pom.xml
@@ -21,7 +21,7 @@
   parent
 groupIdorg.apache.spark/groupId
 artifactIdspark-parent/artifactId
-version1.0.0-SNAPSHOT/version
+version1.0.0/version
 relativePath../pom.xml/relativePath
   /parent
 
@@ -122,8 +122,8 @@
 transformer 
implementation=org.apache.maven.plugins.shade.resource.DontIncludeResourceTransformer
   resourcelog4j.properties/resource
 /transformer
-transformer 
implementation=org.apache.maven.plugins.shade.resource.ApacheLicenseResourceTransformer/
-transformer 
implementation=org.apache.maven.plugins.shade.resource.ApacheNoticeResourceTransformer/
+transformer 
implementation=org.apache.maven.plugins.shade.resource.ApacheLicenseResourceTransformer
 /
+transformer 
implementation=org.apache.maven.plugins.shade.resource.ApacheNoticeResourceTransformer
 /
   /transformers
 /configuration
   /execution

http://git-wip-us.apache.org/repos/asf/spark/blob/2f1dc868/bagel/pom.xml
--
diff --git a/bagel/pom.xml b/bagel/pom.xml
index 355f437..08932bb 100644
--- a/bagel/pom.xml
+++ b/bagel/pom.xml
@@ -21,7 +21,7 @@
   parent
 groupIdorg.apache.spark/groupId
 artifactIdspark-parent/artifactId
-version1.0.0-SNAPSHOT/version
+version1.0.0/version
 relativePath../pom.xml/relativePath
   /parent
 

http://git-wip-us.apache.org/repos/asf/spark/blob/2f1dc868/core/pom.xml
--
diff --git a/core/pom.xml b/core/pom.xml
index bab50f5..3e22641 100644
--- a/core/pom.xml
+++ b/core/pom.xml
@@ -21,7 +21,7 @@
   parent
 groupIdorg.apache.spark/groupId
 artifactIdspark-parent/artifactId
-version1.0.0-SNAPSHOT/version
+version1.0.0/version
 relativePath../pom.xml/relativePath
   /parent
 

http://git-wip-us.apache.org/repos/asf/spark/blob/2f1dc868/examples/pom.xml
--
diff --git a/examples/pom.xml b/examples/pom.xml
index 874bcd7..006757a 100644
--- a/examples/pom.xml
+++ b/examples/pom.xml
@@ -21,7 +21,7 @@
   parent
 groupIdorg.apache.spark/groupId
 artifactIdspark-parent/artifactId
-version1.0.0-SNAPSHOT/version
+version1.0.0/version
 relativePath../pom.xml/relativePath
   /parent
 

http://git-wip-us.apache.org/repos/asf/spark/blob/2f1dc868/external/flume/pom.xml
--
diff --git a/external/flume/pom.xml b/external/flume/pom.xml
index 6aec215..3ba984e 100644
--- a/external/flume/pom.xml
+++ b/external/flume/pom.xml
@@ -21,7 +21,7 @@
   parent
 groupIdorg.apache.spark/groupId
 artifactIdspark-parent/artifactId
-version1.0.0-SNAPSHOT/version
+version1.0.0/version
 relativePath../../pom.xml/relativePath
   /parent
 


Git Push Summary

2014-05-25 Thread tdas
Repository: spark
Updated Tags:  refs/tags/v1.0.0-rc11 [created] 522777f7e


[3/3] git commit: Updated CHANGES.txt

2014-05-26 Thread tdas
Updated CHANGES.txt


Project: http://git-wip-us.apache.org/repos/asf/spark/repo
Commit: http://git-wip-us.apache.org/repos/asf/spark/commit/caed16e4
Tree: http://git-wip-us.apache.org/repos/asf/spark/tree/caed16e4
Diff: http://git-wip-us.apache.org/repos/asf/spark/diff/caed16e4

Branch: refs/heads/branch-1.0
Commit: caed16e4c86427f13b3d4604c5dcb9ddcb71bbb1
Parents: 6d34a6a
Author: Tathagata Das tathagata.das1...@gmail.com
Authored: Mon May 26 06:16:25 2014 +
Committer: Tathagata Das tathagata.das1...@gmail.com
Committed: Mon May 26 06:16:25 2014 +

--
 CHANGES.txt | 65 
 1 file changed, 65 insertions(+)
--


http://git-wip-us.apache.org/repos/asf/spark/blob/caed16e4/CHANGES.txt
--
diff --git a/CHANGES.txt b/CHANGES.txt
index b097339..0cff06c 100644
--- a/CHANGES.txt
+++ b/CHANGES.txt
@@ -3,6 +3,71 @@ Spark Change Log
 
 Release 1.0.0
 
+  HOTFIX: Add no-arg SparkContext constructor in Java
+  Patrick Wendell pwend...@gmail.com
+  2014-05-25 20:13:32 -0700
+  Commit: 18c77cb, github.com/apache/spark/pull/878
+
+  [SQL] Minor: Introduce SchemaRDD#aggregate() for simple aggregations
+  Aaron Davidson aa...@databricks.com
+  2014-05-25 18:37:44 -0700
+  Commit: a3976a2, github.com/apache/spark/pull/874
+
+  SPARK-1903 Document Spark's network connections
+  Andrew Ash and...@andrewash.com
+  2014-05-25 17:15:47 -0700
+  Commit: 5107a6f, github.com/apache/spark/pull/856
+
+  Fix PEP8 violations in Python mllib.
+  Reynold Xin r...@apache.org
+  2014-05-25 17:15:01 -0700
+  Commit: 07f34ca, github.com/apache/spark/pull/871
+
+  Python docstring update for sql.py.
+  Reynold Xin r...@apache.org
+  2014-05-25 16:04:17 -0700
+  Commit: 8891495, github.com/apache/spark/pull/869
+
+  Fix PEP8 violations in examples/src/main/python.
+  Reynold Xin r...@apache.org
+  2014-05-25 14:48:27 -0700
+  Commit: 3368397, github.com/apache/spark/pull/870
+
+  Added license header for tox.ini.
+  Reynold Xin r...@apache.org
+  2014-05-25 01:47:08 -0700
+  Commit: 7273bfc
+
+  SPARK-1822: Some minor cleanup work on SchemaRDD.count()
+  Reynold Xin r...@apache.org
+  2014-05-25 01:44:49 -0700
+  Commit: aeffc20, github.com/apache/spark/pull/868
+
+  Added PEP8 style configuration file.
+  Reynold Xin r...@apache.org
+  2014-05-25 01:32:15 -0700
+  Commit: 291567d, github.com/apache/spark/pull/872
+
+  [SPARK-1822] SchemaRDD.count() should use query optimizer
+  Kan Zhang kzh...@apache.org
+  2014-05-25 00:06:42 -0700
+  Commit: 64d0fb5, github.com/apache/spark/pull/841
+
+  spark-submit: add exec at the end of the script
+  Colin Patrick Mccabe cmcc...@cloudera.com
+  2014-05-24 22:39:27 -0700
+  Commit: 7e59335, github.com/apache/spark/pull/858
+
+  [SPARK-1886] check executor id existence when executor exit
+  Zhen Peng zhenpen...@baidu.com
+  2014-05-24 20:40:19 -0700
+  Commit: b5e9686, github.com/apache/spark/pull/827
+
+  Updated CHANGES.txt
+  Tathagata Das tathagata.das1...@gmail.com
+  2014-05-25 02:20:13 +
+  Commit: 8406092
+
   SPARK-1911: Emphasize that Spark jars should be built with Java 6.
   Patrick Wendell pwend...@gmail.com
   2014-05-24 18:27:00 -0700



[1/3] git commit: Revert [maven-release-plugin] prepare for next development iteration

2014-05-26 Thread tdas
Repository: spark
Updated Branches:
  refs/heads/branch-1.0 18c77cb53 - caed16e4c


Revert [maven-release-plugin] prepare for next development iteration

This reverts commit 832dc594e7666f1d402334f8015ce29917d9c888.


Project: http://git-wip-us.apache.org/repos/asf/spark/repo
Commit: http://git-wip-us.apache.org/repos/asf/spark/commit/73ffd1e5
Tree: http://git-wip-us.apache.org/repos/asf/spark/tree/73ffd1e5
Diff: http://git-wip-us.apache.org/repos/asf/spark/diff/73ffd1e5

Branch: refs/heads/branch-1.0
Commit: 73ffd1e547e14475c23e62402c800d7d18260365
Parents: 18c77cb
Author: Tathagata Das tathagata.das1...@gmail.com
Authored: Mon May 26 06:09:20 2014 +
Committer: Tathagata Das tathagata.das1...@gmail.com
Committed: Mon May 26 06:09:20 2014 +

--
 assembly/pom.xml  | 2 +-
 bagel/pom.xml | 2 +-
 core/pom.xml  | 2 +-
 examples/pom.xml  | 2 +-
 external/flume/pom.xml| 2 +-
 external/kafka/pom.xml| 2 +-
 external/mqtt/pom.xml | 2 +-
 external/twitter/pom.xml  | 2 +-
 external/zeromq/pom.xml   | 2 +-
 extras/spark-ganglia-lgpl/pom.xml | 2 +-
 graphx/pom.xml| 2 +-
 mllib/pom.xml | 2 +-
 pom.xml   | 4 ++--
 repl/pom.xml  | 2 +-
 sql/catalyst/pom.xml  | 2 +-
 sql/core/pom.xml  | 2 +-
 sql/hive/pom.xml  | 2 +-
 streaming/pom.xml | 2 +-
 tools/pom.xml | 2 +-
 yarn/pom.xml  | 2 +-
 yarn/stable/pom.xml   | 2 +-
 21 files changed, 22 insertions(+), 22 deletions(-)
--


http://git-wip-us.apache.org/repos/asf/spark/blob/73ffd1e5/assembly/pom.xml
--
diff --git a/assembly/pom.xml b/assembly/pom.xml
index 720ca77..bed4261 100644
--- a/assembly/pom.xml
+++ b/assembly/pom.xml
@@ -21,7 +21,7 @@
   parent
 groupIdorg.apache.spark/groupId
 artifactIdspark-parent/artifactId
-version1.0.1-SNAPSHOT/version
+version1.0.0/version
 relativePath../pom.xml/relativePath
   /parent
 

http://git-wip-us.apache.org/repos/asf/spark/blob/73ffd1e5/bagel/pom.xml
--
diff --git a/bagel/pom.xml b/bagel/pom.xml
index 85f6d99..08932bb 100644
--- a/bagel/pom.xml
+++ b/bagel/pom.xml
@@ -21,7 +21,7 @@
   parent
 groupIdorg.apache.spark/groupId
 artifactIdspark-parent/artifactId
-version1.0.1-SNAPSHOT/version
+version1.0.0/version
 relativePath../pom.xml/relativePath
   /parent
 

http://git-wip-us.apache.org/repos/asf/spark/blob/73ffd1e5/core/pom.xml
--
diff --git a/core/pom.xml b/core/pom.xml
index 47c2507..3e22641 100644
--- a/core/pom.xml
+++ b/core/pom.xml
@@ -21,7 +21,7 @@
   parent
 groupIdorg.apache.spark/groupId
 artifactIdspark-parent/artifactId
-version1.0.1-SNAPSHOT/version
+version1.0.0/version
 relativePath../pom.xml/relativePath
   /parent
 

http://git-wip-us.apache.org/repos/asf/spark/blob/73ffd1e5/examples/pom.xml
--
diff --git a/examples/pom.xml b/examples/pom.xml
index b7cbb1a..006757a 100644
--- a/examples/pom.xml
+++ b/examples/pom.xml
@@ -21,7 +21,7 @@
   parent
 groupIdorg.apache.spark/groupId
 artifactIdspark-parent/artifactId
-version1.0.1-SNAPSHOT/version
+version1.0.0/version
 relativePath../pom.xml/relativePath
   /parent
 

http://git-wip-us.apache.org/repos/asf/spark/blob/73ffd1e5/external/flume/pom.xml
--
diff --git a/external/flume/pom.xml b/external/flume/pom.xml
index b8fc07f..3ba984e 100644
--- a/external/flume/pom.xml
+++ b/external/flume/pom.xml
@@ -21,7 +21,7 @@
   parent
 groupIdorg.apache.spark/groupId
 artifactIdspark-parent/artifactId
-version1.0.1-SNAPSHOT/version
+version1.0.0/version
 relativePath../../pom.xml/relativePath
   /parent
 

http://git-wip-us.apache.org/repos/asf/spark/blob/73ffd1e5/external/kafka/pom.xml
--
diff --git a/external/kafka/pom.xml b/external/kafka/pom.xml
index 9eeb2e1..cb4dd47 100644
--- a/external/kafka/pom.xml
+++ b/external/kafka/pom.xml
@@ -21,7 +21,7 @@
   parent
 groupIdorg.apache.spark/groupId
 artifactIdspark-parent/artifactId
-version1.0.1-SNAPSHOT/version
+version1.0.0/version
 relativePath../../pom.xml/relativePath
   /parent
 

http://git-wip-us.apache.org/repos/asf/spark/blob/73ffd1e5/external/mqtt/pom.xml
--
diff --git 

Git Push Summary

2014-05-26 Thread tdas
Repository: spark
Updated Tags:  refs/tags/v1.0.0-rc11 [deleted] 522777f7e


git commit: [maven-release-plugin] prepare release v1.0.0-rc11

2014-05-26 Thread tdas
Repository: spark
Updated Branches:
  refs/heads/branch-1.0 caed16e4c - c69d97cdb


[maven-release-plugin] prepare release v1.0.0-rc11


Project: http://git-wip-us.apache.org/repos/asf/spark/repo
Commit: http://git-wip-us.apache.org/repos/asf/spark/commit/c69d97cd
Tree: http://git-wip-us.apache.org/repos/asf/spark/tree/c69d97cd
Diff: http://git-wip-us.apache.org/repos/asf/spark/diff/c69d97cd

Branch: refs/heads/branch-1.0
Commit: c69d97cdb42f809cb71113a1db4194c21372242a
Parents: caed16e
Author: Tathagata Das tathagata.das1...@gmail.com
Authored: Mon May 26 06:46:48 2014 +
Committer: Tathagata Das tathagata.das1...@gmail.com
Committed: Mon May 26 06:46:48 2014 +

--
 assembly/pom.xml  | 6 +++---
 bagel/pom.xml | 2 +-
 core/pom.xml  | 2 +-
 examples/pom.xml  | 2 +-
 external/flume/pom.xml| 2 +-
 external/kafka/pom.xml| 2 +-
 external/mqtt/pom.xml | 2 +-
 external/twitter/pom.xml  | 2 +-
 external/zeromq/pom.xml   | 2 +-
 extras/spark-ganglia-lgpl/pom.xml | 2 +-
 graphx/pom.xml| 2 +-
 mllib/pom.xml | 2 +-
 pom.xml   | 4 ++--
 repl/pom.xml  | 2 +-
 sql/catalyst/pom.xml  | 2 +-
 sql/core/pom.xml  | 2 +-
 sql/hive/pom.xml  | 2 +-
 streaming/pom.xml | 2 +-
 tools/pom.xml | 2 +-
 yarn/pom.xml  | 2 +-
 yarn/stable/pom.xml   | 2 +-
 21 files changed, 24 insertions(+), 24 deletions(-)
--


http://git-wip-us.apache.org/repos/asf/spark/blob/c69d97cd/assembly/pom.xml
--
diff --git a/assembly/pom.xml b/assembly/pom.xml
index 963357b..bed4261 100644
--- a/assembly/pom.xml
+++ b/assembly/pom.xml
@@ -21,7 +21,7 @@
   parent
 groupIdorg.apache.spark/groupId
 artifactIdspark-parent/artifactId
-version1.0.0-SNAPSHOT/version
+version1.0.0/version
 relativePath../pom.xml/relativePath
   /parent
 
@@ -122,8 +122,8 @@
 transformer 
implementation=org.apache.maven.plugins.shade.resource.DontIncludeResourceTransformer
   resourcelog4j.properties/resource
 /transformer
-transformer 
implementation=org.apache.maven.plugins.shade.resource.ApacheLicenseResourceTransformer/
-transformer 
implementation=org.apache.maven.plugins.shade.resource.ApacheNoticeResourceTransformer/
+transformer 
implementation=org.apache.maven.plugins.shade.resource.ApacheLicenseResourceTransformer
 /
+transformer 
implementation=org.apache.maven.plugins.shade.resource.ApacheNoticeResourceTransformer
 /
   /transformers
 /configuration
   /execution

http://git-wip-us.apache.org/repos/asf/spark/blob/c69d97cd/bagel/pom.xml
--
diff --git a/bagel/pom.xml b/bagel/pom.xml
index 355f437..08932bb 100644
--- a/bagel/pom.xml
+++ b/bagel/pom.xml
@@ -21,7 +21,7 @@
   parent
 groupIdorg.apache.spark/groupId
 artifactIdspark-parent/artifactId
-version1.0.0-SNAPSHOT/version
+version1.0.0/version
 relativePath../pom.xml/relativePath
   /parent
 

http://git-wip-us.apache.org/repos/asf/spark/blob/c69d97cd/core/pom.xml
--
diff --git a/core/pom.xml b/core/pom.xml
index bab50f5..3e22641 100644
--- a/core/pom.xml
+++ b/core/pom.xml
@@ -21,7 +21,7 @@
   parent
 groupIdorg.apache.spark/groupId
 artifactIdspark-parent/artifactId
-version1.0.0-SNAPSHOT/version
+version1.0.0/version
 relativePath../pom.xml/relativePath
   /parent
 

http://git-wip-us.apache.org/repos/asf/spark/blob/c69d97cd/examples/pom.xml
--
diff --git a/examples/pom.xml b/examples/pom.xml
index 874bcd7..006757a 100644
--- a/examples/pom.xml
+++ b/examples/pom.xml
@@ -21,7 +21,7 @@
   parent
 groupIdorg.apache.spark/groupId
 artifactIdspark-parent/artifactId
-version1.0.0-SNAPSHOT/version
+version1.0.0/version
 relativePath../pom.xml/relativePath
   /parent
 

http://git-wip-us.apache.org/repos/asf/spark/blob/c69d97cd/external/flume/pom.xml
--
diff --git a/external/flume/pom.xml b/external/flume/pom.xml
index 6aec215..3ba984e 100644
--- a/external/flume/pom.xml
+++ b/external/flume/pom.xml
@@ -21,7 +21,7 @@
   parent
 groupIdorg.apache.spark/groupId
 artifactIdspark-parent/artifactId
-version1.0.0-SNAPSHOT/version
+version1.0.0/version
 relativePath../../pom.xml/relativePath
   /parent
 


git commit: Streaming programming guide typos

2014-07-03 Thread tdas
Repository: spark
Updated Branches:
  refs/heads/master d4c30cd99 - fdc4c112e


Streaming programming guide typos

Fix a bad Java code sample and a broken link in the streaming programming guide.

Author: Clément MATHIEU clem...@unportant.info

Closes #1286 from cykl/streaming-programming-guide-typos and squashes the 
following commits:

b0908cb [Clément MATHIEU] Fix broken URL
9d3c535 [Clément MATHIEU] Spark streaming requires at least two working 
threads (scala version was OK)


Project: http://git-wip-us.apache.org/repos/asf/spark/repo
Commit: http://git-wip-us.apache.org/repos/asf/spark/commit/fdc4c112
Tree: http://git-wip-us.apache.org/repos/asf/spark/tree/fdc4c112
Diff: http://git-wip-us.apache.org/repos/asf/spark/diff/fdc4c112

Branch: refs/heads/master
Commit: fdc4c112e7c2ac585d108d03209a642aa8bab7c8
Parents: d4c30cd
Author: Clément MATHIEU clem...@unportant.info
Authored: Thu Jul 3 18:31:18 2014 -0700
Committer: Tathagata Das tathagata.das1...@gmail.com
Committed: Thu Jul 3 18:31:18 2014 -0700

--
 docs/streaming-programming-guide.md | 4 ++--
 1 file changed, 2 insertions(+), 2 deletions(-)
--


http://git-wip-us.apache.org/repos/asf/spark/blob/fdc4c112/docs/streaming-programming-guide.md
--
diff --git a/docs/streaming-programming-guide.md 
b/docs/streaming-programming-guide.md
index ce8e58d..90a0eef 100644
--- a/docs/streaming-programming-guide.md
+++ b/docs/streaming-programming-guide.md
@@ -148,7 +148,7 @@ import org.apache.spark.streaming.*;
 import org.apache.spark.streaming.api.java.*;
 import scala.Tuple2;
 // Create a StreamingContext with a local master
-JavaStreamingContext jssc = new JavaStreamingContext(local, 
JavaNetworkWordCount, new Duration(1000))
+JavaStreamingContext jssc = new JavaStreamingContext(local[2], 
JavaNetworkWordCount, new Duration(1000))
 {% endhighlight %}
 
 Using this context, we then create a new DStream
@@ -216,7 +216,7 @@ jssc.awaitTermination();   // Wait for the computation to 
terminate
 {% endhighlight %}
 
 The complete code can be found in the Spark Streaming example
-[JavaNetworkWordCount]({{site.SPARK_GITHUB_URL}}/blob/master/examples/src/main/java/index.html?org/apache/spark/examples/streaming/JavaNetworkWordCount.java).
+[JavaNetworkWordCount]({{site.SPARK_GITHUB_URL}}/blob/master/examples/src/main/java/org/apache/spark/examples/streaming/JavaNetworkWordCount.java).
 br
 
 /div



git commit: [SPARK-2362] Fix for newFilesOnly logic in file DStream

2014-07-08 Thread tdas
Repository: spark
Updated Branches:
  refs/heads/branch-1.0 1c12b0b5c - 885489112


[SPARK-2362] Fix for newFilesOnly logic in file DStream

The newFilesOnly logic should be inverted: the logic should be that if the flag 
newFilesOnly==true then only start reading files older than current time. As 
the code is now if newFilesOnly==true then it will start to read files that are 
older than 0L (that is: every file in the directory).

Author: Gabriele Nizzoli m...@nizzoli.net

Closes #1077 from gabrielenizzoli/master and squashes the following commits:

4f1d261 [Gabriele Nizzoli] Fix for newFilesOnly logic in file DStream

(cherry picked from commit e6f7bfcfbf6aff7a9f8cd8e0a2166d0bf62b0912)
Signed-off-by: Tathagata Das tathagata.das1...@gmail.com


Project: http://git-wip-us.apache.org/repos/asf/spark/repo
Commit: http://git-wip-us.apache.org/repos/asf/spark/commit/88548911
Tree: http://git-wip-us.apache.org/repos/asf/spark/tree/88548911
Diff: http://git-wip-us.apache.org/repos/asf/spark/diff/88548911

Branch: refs/heads/branch-1.0
Commit: 885489112c82eb909df7efbf0515fd7abfae41a4
Parents: 1c12b0b
Author: Gabriele Nizzoli m...@nizzoli.net
Authored: Tue Jul 8 14:23:38 2014 -0700
Committer: Tathagata Das tathagata.das1...@gmail.com
Committed: Tue Jul 8 14:24:50 2014 -0700

--
 .../org/apache/spark/streaming/dstream/FileInputDStream.scala  | 2 +-
 1 file changed, 1 insertion(+), 1 deletion(-)
--


http://git-wip-us.apache.org/repos/asf/spark/blob/88548911/streaming/src/main/scala/org/apache/spark/streaming/dstream/FileInputDStream.scala
--
diff --git 
a/streaming/src/main/scala/org/apache/spark/streaming/dstream/FileInputDStream.scala
 
b/streaming/src/main/scala/org/apache/spark/streaming/dstream/FileInputDStream.scala
index e878285..9eecbfa 100644
--- 
a/streaming/src/main/scala/org/apache/spark/streaming/dstream/FileInputDStream.scala
+++ 
b/streaming/src/main/scala/org/apache/spark/streaming/dstream/FileInputDStream.scala
@@ -45,7 +45,7 @@ class FileInputDStream[K: ClassTag, V: ClassTag, F : 
NewInputFormat[K,V] : Clas
   // Files with mod time earlier than this is ignored. This is updated every 
interval
   // such that in the current interval, files older than any file found in the
   // previous interval will be ignored. Obviously this time keeps moving 
forward.
-  private var ignoreTime = if (newFilesOnly) 0L else System.currentTimeMillis()
+  private var ignoreTime = if (newFilesOnly) System.currentTimeMillis() else 0L
 
   // Latest file mod time seen till any point of time
   @transient private var path_ : Path = null



git commit: [STREAMING] SPARK-2343: Fix QueueInputDStream with oneAtATime false

2014-07-09 Thread tdas
Repository: spark
Updated Branches:
  refs/heads/branch-1.0 d569838bc - 21fae6d3e


[STREAMING] SPARK-2343: Fix QueueInputDStream with oneAtATime false

Fix QueueInputDStream which was not removing dequeued items when used with the 
oneAtATime flag disabled.

Author: Manuel Laflamme manuel.lafla...@gmail.com

Closes #1285 from mlaflamm/spark-2343 and squashes the following commits:

61c9e38 [Manuel Laflamme] Unit tests for queue input stream
c51d029 [Manuel Laflamme] Fix QueueInputDStream with oneAtATime false

(cherry picked from commit 0eb11527d13083ced215e3fda44ed849198a57cb)
Signed-off-by: Tathagata Das tathagata.das1...@gmail.com


Project: http://git-wip-us.apache.org/repos/asf/spark/repo
Commit: http://git-wip-us.apache.org/repos/asf/spark/commit/21fae6d3
Tree: http://git-wip-us.apache.org/repos/asf/spark/tree/21fae6d3
Diff: http://git-wip-us.apache.org/repos/asf/spark/diff/21fae6d3

Branch: refs/heads/branch-1.0
Commit: 21fae6d3e6f96d7272b55b8ee2ed2571c558b1fe
Parents: d569838
Author: Manuel Laflamme manuel.lafla...@gmail.com
Authored: Wed Jul 9 10:45:45 2014 -0700
Committer: Tathagata Das tathagata.das1...@gmail.com
Committed: Wed Jul 9 10:46:12 2014 -0700

--
 .../streaming/dstream/QueueInputDStream.scala   |  2 +-
 .../spark/streaming/InputStreamsSuite.scala | 92 +++-
 2 files changed, 92 insertions(+), 2 deletions(-)
--


http://git-wip-us.apache.org/repos/asf/spark/blob/21fae6d3/streaming/src/main/scala/org/apache/spark/streaming/dstream/QueueInputDStream.scala
--
diff --git 
a/streaming/src/main/scala/org/apache/spark/streaming/dstream/QueueInputDStream.scala
 
b/streaming/src/main/scala/org/apache/spark/streaming/dstream/QueueInputDStream.scala
index 6376cff..ed7da6d 100644
--- 
a/streaming/src/main/scala/org/apache/spark/streaming/dstream/QueueInputDStream.scala
+++ 
b/streaming/src/main/scala/org/apache/spark/streaming/dstream/QueueInputDStream.scala
@@ -41,7 +41,7 @@ class QueueInputDStream[T: ClassTag](
 if (oneAtATime  queue.size  0) {
   buffer += queue.dequeue()
 } else {
-  buffer ++= queue
+  buffer ++= queue.dequeueAll(_ = true)
 }
 if (buffer.size  0) {
   if (oneAtATime) {

http://git-wip-us.apache.org/repos/asf/spark/blob/21fae6d3/streaming/src/test/scala/org/apache/spark/streaming/InputStreamsSuite.scala
--
diff --git 
a/streaming/src/test/scala/org/apache/spark/streaming/InputStreamsSuite.scala 
b/streaming/src/test/scala/org/apache/spark/streaming/InputStreamsSuite.scala
index cd0aa4d..cc4a650 100644
--- 
a/streaming/src/test/scala/org/apache/spark/streaming/InputStreamsSuite.scala
+++ 
b/streaming/src/test/scala/org/apache/spark/streaming/InputStreamsSuite.scala
@@ -29,7 +29,7 @@ import java.nio.charset.Charset
 import java.util.concurrent.{Executors, TimeUnit, ArrayBlockingQueue}
 import java.util.concurrent.atomic.AtomicInteger
 
-import scala.collection.mutable.{SynchronizedBuffer, ArrayBuffer}
+import scala.collection.mutable.{SynchronizedBuffer, ArrayBuffer, 
SynchronizedQueue}
 
 import com.google.common.io.Files
 import org.scalatest.BeforeAndAfter
@@ -39,6 +39,7 @@ import org.apache.spark.storage.StorageLevel
 import org.apache.spark.streaming.util.ManualClock
 import org.apache.spark.util.Utils
 import org.apache.spark.streaming.receiver.{ActorHelper, Receiver}
+import org.apache.spark.rdd.RDD
 
 class InputStreamsSuite extends TestSuiteBase with BeforeAndAfter {
 
@@ -234,6 +235,95 @@ class InputStreamsSuite extends TestSuiteBase with 
BeforeAndAfter {
 logInfo()
 assert(output.sum === numTotalRecords)
   }
+
+  test(queue input stream - oneAtATime=true) {
+// Set up the streaming context and input streams
+val ssc = new StreamingContext(conf, batchDuration)
+val queue = new SynchronizedQueue[RDD[String]]()
+val queueStream = ssc.queueStream(queue, oneAtATime = true)
+val outputBuffer = new ArrayBuffer[Seq[String]] with 
SynchronizedBuffer[Seq[String]]
+val outputStream = new TestOutputStream(queueStream, outputBuffer)
+def output = outputBuffer.filter(_.size  0)
+outputStream.register()
+ssc.start()
+
+// Setup data queued into the stream
+val clock = ssc.scheduler.clock.asInstanceOf[ManualClock]
+val input = Seq(1, 2, 3, 4, 5)
+val expectedOutput = input.map(Seq(_))
+//Thread.sleep(1000)
+val inputIterator = input.toIterator
+for (i - 0 until input.size) {
+  // Enqueue more than 1 item per tick but they should dequeue one at a 
time
+  inputIterator.take(2).foreach(i = queue += 
ssc.sparkContext.makeRDD(Seq(i)))
+  clock.addToTime(batchDuration.milliseconds)
+}
+Thread.sleep(1000)
+logInfo(Stopping context)
+

git commit: [SPARK-1478].3: Upgrade FlumeInputDStream's FlumeReceiver to support FLUME-1915

2014-07-10 Thread tdas
Repository: spark
Updated Branches:
  refs/heads/master 369aa84e8 - 40a8fef4e


[SPARK-1478].3: Upgrade FlumeInputDStream's FlumeReceiver to support FLUME-1915

This is a modified version of this PR https://github.com/apache/spark/pull/1168 
done by @tmalaska
Adds MIMA binary check exclusions.

Author: tmalaska ted.mala...@cloudera.com
Author: Tathagata Das tathagata.das1...@gmail.com

Closes #1347 from tdas/FLUME-1915 and squashes the following commits:

96065df [Tathagata Das] Added Mima exclusion for FlumeReceiver.
41d5338 [tmalaska] Address line 57 that was too long
12617e5 [tmalaska] SPARK-1478: Upgrade FlumeInputDStream's Flume...


Project: http://git-wip-us.apache.org/repos/asf/spark/repo
Commit: http://git-wip-us.apache.org/repos/asf/spark/commit/40a8fef4
Tree: http://git-wip-us.apache.org/repos/asf/spark/tree/40a8fef4
Diff: http://git-wip-us.apache.org/repos/asf/spark/diff/40a8fef4

Branch: refs/heads/master
Commit: 40a8fef4e6619b4ea10a4ec9026260649ce5ae73
Parents: 369aa84
Author: tmalaska ted.mala...@cloudera.com
Authored: Thu Jul 10 13:15:02 2014 -0700
Committer: Tathagata Das tathagata.das1...@gmail.com
Committed: Thu Jul 10 13:15:02 2014 -0700

--
 .../streaming/flume/FlumeInputDStream.scala | 76 +---
 .../spark/streaming/flume/FlumeUtils.scala  | 41 ++-
 .../streaming/flume/JavaFlumeStreamSuite.java   |  2 +
 .../streaming/flume/FlumeStreamSuite.scala  | 41 +--
 project/MimaExcludes.scala  |  3 +
 5 files changed, 147 insertions(+), 16 deletions(-)
--


http://git-wip-us.apache.org/repos/asf/spark/blob/40a8fef4/external/flume/src/main/scala/org/apache/spark/streaming/flume/FlumeInputDStream.scala
--
diff --git 
a/external/flume/src/main/scala/org/apache/spark/streaming/flume/FlumeInputDStream.scala
 
b/external/flume/src/main/scala/org/apache/spark/streaming/flume/FlumeInputDStream.scala
index ed35e34..07ae88f 100644
--- 
a/external/flume/src/main/scala/org/apache/spark/streaming/flume/FlumeInputDStream.scala
+++ 
b/external/flume/src/main/scala/org/apache/spark/streaming/flume/FlumeInputDStream.scala
@@ -20,6 +20,7 @@ package org.apache.spark.streaming.flume
 import java.net.InetSocketAddress
 import java.io.{ObjectInput, ObjectOutput, Externalizable}
 import java.nio.ByteBuffer
+import java.util.concurrent.Executors
 
 import scala.collection.JavaConversions._
 import scala.reflect.ClassTag
@@ -29,24 +30,32 @@ import org.apache.flume.source.avro.AvroFlumeEvent
 import org.apache.flume.source.avro.Status
 import org.apache.avro.ipc.specific.SpecificResponder
 import org.apache.avro.ipc.NettyServer
-
+import org.apache.spark.Logging
 import org.apache.spark.util.Utils
 import org.apache.spark.storage.StorageLevel
-import org.apache.spark.streaming.StreamingContext
 import org.apache.spark.streaming.dstream._
-import org.apache.spark.Logging
+import org.apache.spark.streaming.StreamingContext
 import org.apache.spark.streaming.receiver.Receiver
 
+import org.jboss.netty.channel.ChannelPipelineFactory
+import org.jboss.netty.channel.Channels
+import org.jboss.netty.channel.ChannelPipeline
+import org.jboss.netty.channel.ChannelFactory
+import org.jboss.netty.channel.socket.nio.NioServerSocketChannelFactory
+import org.jboss.netty.handler.codec.compression._
+import org.jboss.netty.handler.execution.ExecutionHandler
+
 private[streaming]
 class FlumeInputDStream[T: ClassTag](
   @transient ssc_ : StreamingContext,
   host: String,
   port: Int,
-  storageLevel: StorageLevel
+  storageLevel: StorageLevel,
+  enableDecompression: Boolean
 ) extends ReceiverInputDStream[SparkFlumeEvent](ssc_) {
 
   override def getReceiver(): Receiver[SparkFlumeEvent] = {
-new FlumeReceiver(host, port, storageLevel)
+new FlumeReceiver(host, port, storageLevel, enableDecompression)
   }
 }
 
@@ -134,22 +143,71 @@ private[streaming]
 class FlumeReceiver(
 host: String,
 port: Int,
-storageLevel: StorageLevel
+storageLevel: StorageLevel,
+enableDecompression: Boolean
   ) extends Receiver[SparkFlumeEvent](storageLevel) with Logging {
 
   lazy val responder = new SpecificResponder(
 classOf[AvroSourceProtocol], new FlumeEventServer(this))
-  lazy val server = new NettyServer(responder, new InetSocketAddress(host, 
port))
+  var server: NettyServer = null
+
+  private def initServer() = {
+if (enableDecompression) {
+  val channelFactory = new NioServerSocketChannelFactory
+(Executors.newCachedThreadPool(), Executors.newCachedThreadPool());
+  val channelPipelieFactory = new CompressionChannelPipelineFactory()
+  
+  new NettyServer(
+responder, 
+new InetSocketAddress(host, port),
+channelFactory, 
+channelPipelieFactory, 
+null)
+} else {
+  new

git commit: [SPARK-1341] [Streaming] Throttle BlockGenerator to limit rate of data consumption.

2014-07-10 Thread tdas
Repository: spark
Updated Branches:
  refs/heads/master 40a8fef4e - 2dd672485


[SPARK-1341] [Streaming] Throttle BlockGenerator to limit rate of data 
consumption.

Author: Issac Buenrostro buenros...@ooyala.com

Closes #945 from ibuenros/SPARK-1341-throttle and squashes the following 
commits:

5514916 [Issac Buenrostro] Formatting changes, added documentation for 
streaming throttling, stricter unit tests for throttling.
62f395f [Issac Buenrostro] Add comments and license to streaming 
RateLimiter.scala
7066438 [Issac Buenrostro] Moved throttle code to RateLimiter class, smoother 
pushing when throttling active
ccafe09 [Issac Buenrostro] Throttle BlockGenerator to limit rate of data 
consumption.


Project: http://git-wip-us.apache.org/repos/asf/spark/repo
Commit: http://git-wip-us.apache.org/repos/asf/spark/commit/2dd67248
Tree: http://git-wip-us.apache.org/repos/asf/spark/tree/2dd67248
Diff: http://git-wip-us.apache.org/repos/asf/spark/diff/2dd67248

Branch: refs/heads/master
Commit: 2dd67248503306bb08946b1796821e9f9ed4d00e
Parents: 40a8fef
Author: Issac Buenrostro buenros...@ooyala.com
Authored: Thu Jul 10 16:01:08 2014 -0700
Committer: Tathagata Das tathagata.das1...@gmail.com
Committed: Thu Jul 10 16:01:08 2014 -0700

--
 docs/configuration.md   |  9 +++
 .../streaming/receiver/BlockGenerator.scala |  3 +-
 .../spark/streaming/receiver/RateLimiter.scala  | 69 
 .../spark/streaming/NetworkReceiverSuite.scala  | 38 +++
 4 files changed, 118 insertions(+), 1 deletion(-)
--


http://git-wip-us.apache.org/repos/asf/spark/blob/2dd67248/docs/configuration.md
--
diff --git a/docs/configuration.md b/docs/configuration.md
index b84104c..0aea23a 100644
--- a/docs/configuration.md
+++ b/docs/configuration.md
@@ -774,6 +774,15 @@ Apart from these, the following properties are also 
available, and may be useful
   /td
 /tr
 tr
+  tdcodespark.streaming.receiver.maxRate/code/td
+  tdinfinite/td
+  td
+Maximum rate (per second) at which each receiver will push data into 
blocks. Effectively,
+each stream will consume at most this number of records per second.
+Setting this configuration to 0 or a negative number will put no limit on 
the rate.
+  /td
+/tr
+tr
   tdcodespark.streaming.unpersist/code/td
   tdtrue/td
   td

http://git-wip-us.apache.org/repos/asf/spark/blob/2dd67248/streaming/src/main/scala/org/apache/spark/streaming/receiver/BlockGenerator.scala
--
diff --git 
a/streaming/src/main/scala/org/apache/spark/streaming/receiver/BlockGenerator.scala
 
b/streaming/src/main/scala/org/apache/spark/streaming/receiver/BlockGenerator.scala
index 78cc2da..0316b68 100644
--- 
a/streaming/src/main/scala/org/apache/spark/streaming/receiver/BlockGenerator.scala
+++ 
b/streaming/src/main/scala/org/apache/spark/streaming/receiver/BlockGenerator.scala
@@ -44,7 +44,7 @@ private[streaming] class BlockGenerator(
 listener: BlockGeneratorListener,
 receiverId: Int,
 conf: SparkConf
-  ) extends Logging {
+  ) extends RateLimiter(conf) with Logging {
 
   private case class Block(id: StreamBlockId, buffer: ArrayBuffer[Any])
 
@@ -81,6 +81,7 @@ private[streaming] class BlockGenerator(
* will be periodically pushed into BlockManager.
*/
   def += (data: Any): Unit = synchronized {
+waitToPush()
 currentBuffer += data
   }
 

http://git-wip-us.apache.org/repos/asf/spark/blob/2dd67248/streaming/src/main/scala/org/apache/spark/streaming/receiver/RateLimiter.scala
--
diff --git 
a/streaming/src/main/scala/org/apache/spark/streaming/receiver/RateLimiter.scala
 
b/streaming/src/main/scala/org/apache/spark/streaming/receiver/RateLimiter.scala
new file mode 100644
index 000..e4f6ba6
--- /dev/null
+++ 
b/streaming/src/main/scala/org/apache/spark/streaming/receiver/RateLimiter.scala
@@ -0,0 +1,69 @@
+/*
+ * Licensed to the Apache Software Foundation (ASF) under one or more
+ * contributor license agreements.  See the NOTICE file distributed with
+ * this work for additional information regarding copyright ownership.
+ * The ASF licenses this file to You under the Apache License, Version 2.0
+ * (the License); you may not use this file except in compliance with
+ * the License.  You may obtain a copy of the License at
+ *
+ *http://www.apache.org/licenses/LICENSE-2.0
+ *
+ * Unless required by applicable law or agreed to in writing, software
+ * distributed under the License is distributed on an AS IS BASIS,
+ * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+ * See the License for the specific language governing permissions and
+ * limitations under the License.
+ */
+

git commit: [SPARK-1341] [Streaming] Throttle BlockGenerator to limit rate of data consumption.

2014-07-10 Thread tdas
Repository: spark
Updated Branches:
  refs/heads/branch-1.0 cb443cf6c - ca19cfbcd


[SPARK-1341] [Streaming] Throttle BlockGenerator to limit rate of data 
consumption.

Author: Issac Buenrostro buenros...@ooyala.com

Closes #945 from ibuenros/SPARK-1341-throttle and squashes the following 
commits:

5514916 [Issac Buenrostro] Formatting changes, added documentation for 
streaming throttling, stricter unit tests for throttling.
62f395f [Issac Buenrostro] Add comments and license to streaming 
RateLimiter.scala
7066438 [Issac Buenrostro] Moved throttle code to RateLimiter class, smoother 
pushing when throttling active
ccafe09 [Issac Buenrostro] Throttle BlockGenerator to limit rate of data 
consumption.

(cherry picked from commit 2dd67248503306bb08946b1796821e9f9ed4d00e)
Signed-off-by: Tathagata Das tathagata.das1...@gmail.com


Project: http://git-wip-us.apache.org/repos/asf/spark/repo
Commit: http://git-wip-us.apache.org/repos/asf/spark/commit/ca19cfbc
Tree: http://git-wip-us.apache.org/repos/asf/spark/tree/ca19cfbc
Diff: http://git-wip-us.apache.org/repos/asf/spark/diff/ca19cfbc

Branch: refs/heads/branch-1.0
Commit: ca19cfbcd5cfac9ad731350dfeea14355aec87d6
Parents: cb443cf
Author: Issac Buenrostro buenros...@ooyala.com
Authored: Thu Jul 10 16:01:08 2014 -0700
Committer: Tathagata Das tathagata.das1...@gmail.com
Committed: Thu Jul 10 16:01:28 2014 -0700

--
 docs/configuration.md   |  9 +++
 .../streaming/receiver/BlockGenerator.scala |  3 +-
 .../spark/streaming/receiver/RateLimiter.scala  | 69 
 .../spark/streaming/NetworkReceiverSuite.scala  | 38 +++
 4 files changed, 118 insertions(+), 1 deletion(-)
--


http://git-wip-us.apache.org/repos/asf/spark/blob/ca19cfbc/docs/configuration.md
--
diff --git a/docs/configuration.md b/docs/configuration.md
index 71fafa5..fad0daf 100644
--- a/docs/configuration.md
+++ b/docs/configuration.md
@@ -774,6 +774,15 @@ Apart from these, the following properties are also 
available, and may be useful
   /td
 /tr
 tr
+  tdcodespark.streaming.receiver.maxRate/code/td
+  tdinfinite/td
+  td
+Maximum rate (per second) at which each receiver will push data into 
blocks. Effectively,
+each stream will consume at most this number of records per second.
+Setting this configuration to 0 or a negative number will put no limit on 
the rate.
+  /td
+/tr
+tr
   tdcodespark.streaming.unpersist/code/td
   tdtrue/td
   td

http://git-wip-us.apache.org/repos/asf/spark/blob/ca19cfbc/streaming/src/main/scala/org/apache/spark/streaming/receiver/BlockGenerator.scala
--
diff --git 
a/streaming/src/main/scala/org/apache/spark/streaming/receiver/BlockGenerator.scala
 
b/streaming/src/main/scala/org/apache/spark/streaming/receiver/BlockGenerator.scala
index 78cc2da..0316b68 100644
--- 
a/streaming/src/main/scala/org/apache/spark/streaming/receiver/BlockGenerator.scala
+++ 
b/streaming/src/main/scala/org/apache/spark/streaming/receiver/BlockGenerator.scala
@@ -44,7 +44,7 @@ private[streaming] class BlockGenerator(
 listener: BlockGeneratorListener,
 receiverId: Int,
 conf: SparkConf
-  ) extends Logging {
+  ) extends RateLimiter(conf) with Logging {
 
   private case class Block(id: StreamBlockId, buffer: ArrayBuffer[Any])
 
@@ -81,6 +81,7 @@ private[streaming] class BlockGenerator(
* will be periodically pushed into BlockManager.
*/
   def += (data: Any): Unit = synchronized {
+waitToPush()
 currentBuffer += data
   }
 

http://git-wip-us.apache.org/repos/asf/spark/blob/ca19cfbc/streaming/src/main/scala/org/apache/spark/streaming/receiver/RateLimiter.scala
--
diff --git 
a/streaming/src/main/scala/org/apache/spark/streaming/receiver/RateLimiter.scala
 
b/streaming/src/main/scala/org/apache/spark/streaming/receiver/RateLimiter.scala
new file mode 100644
index 000..e4f6ba6
--- /dev/null
+++ 
b/streaming/src/main/scala/org/apache/spark/streaming/receiver/RateLimiter.scala
@@ -0,0 +1,69 @@
+/*
+ * Licensed to the Apache Software Foundation (ASF) under one or more
+ * contributor license agreements.  See the NOTICE file distributed with
+ * this work for additional information regarding copyright ownership.
+ * The ASF licenses this file to You under the Apache License, Version 2.0
+ * (the License); you may not use this file except in compliance with
+ * the License.  You may obtain a copy of the License at
+ *
+ *http://www.apache.org/licenses/LICENSE-2.0
+ *
+ * Unless required by applicable law or agreed to in writing, software
+ * distributed under the License is distributed on an AS IS BASIS,
+ * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either 

git commit: SPARK-1478.2 Fix incorrect NioServerSocketChannelFactory constructor call

2014-07-17 Thread tdas
Repository: spark
Updated Branches:
  refs/heads/master d988d345d - 1fcd5dcdd


SPARK-1478.2 Fix incorrect NioServerSocketChannelFactory constructor call

The line break inadvertently means this was interpreted as a call to the no-arg 
constructor. This doesn't exist in older Netty even. (Also fixed a val name 
typo.)

Author: Sean Owen sro...@gmail.com

Closes #1466 from srowen/SPARK-1478.2 and squashes the following commits:

59c3501 [Sean Owen] Line break caused Scala to interpret 
NioServerSocketChannelFactory constructor as the no-arg version, which is not 
even present in some versions of Netty


Project: http://git-wip-us.apache.org/repos/asf/spark/repo
Commit: http://git-wip-us.apache.org/repos/asf/spark/commit/1fcd5dcd
Tree: http://git-wip-us.apache.org/repos/asf/spark/tree/1fcd5dcd
Diff: http://git-wip-us.apache.org/repos/asf/spark/diff/1fcd5dcd

Branch: refs/heads/master
Commit: 1fcd5dcdd8edb0e6989278c95e7f2c7d86c4efb2
Parents: d988d34
Author: Sean Owen sro...@gmail.com
Authored: Thu Jul 17 12:20:48 2014 -0700
Committer: Tathagata Das tathagata.das1...@gmail.com
Committed: Thu Jul 17 12:20:48 2014 -0700

--
 .../apache/spark/streaming/flume/FlumeInputDStream.scala  | 10 +-
 1 file changed, 5 insertions(+), 5 deletions(-)
--


http://git-wip-us.apache.org/repos/asf/spark/blob/1fcd5dcd/external/flume/src/main/scala/org/apache/spark/streaming/flume/FlumeInputDStream.scala
--
diff --git 
a/external/flume/src/main/scala/org/apache/spark/streaming/flume/FlumeInputDStream.scala
 
b/external/flume/src/main/scala/org/apache/spark/streaming/flume/FlumeInputDStream.scala
index 07ae88f..56d2886 100644
--- 
a/external/flume/src/main/scala/org/apache/spark/streaming/flume/FlumeInputDStream.scala
+++ 
b/external/flume/src/main/scala/org/apache/spark/streaming/flume/FlumeInputDStream.scala
@@ -153,15 +153,15 @@ class FlumeReceiver(
 
   private def initServer() = {
 if (enableDecompression) {
-  val channelFactory = new NioServerSocketChannelFactory
-(Executors.newCachedThreadPool(), Executors.newCachedThreadPool());
-  val channelPipelieFactory = new CompressionChannelPipelineFactory()
+  val channelFactory = new 
NioServerSocketChannelFactory(Executors.newCachedThreadPool(),
+ 
Executors.newCachedThreadPool())
+  val channelPipelineFactory = new CompressionChannelPipelineFactory()
   
   new NettyServer(
 responder, 
 new InetSocketAddress(host, port),
-channelFactory, 
-channelPipelieFactory, 
+channelFactory,
+channelPipelineFactory,
 null)
 } else {
   new NettyServer(responder, new InetSocketAddress(host, port))



git commit: [SPARK-2464][Streaming] Fixed Twitter stream stopping bug

2014-07-24 Thread tdas
Repository: spark
Updated Branches:
  refs/heads/master fec641b84 - a45d5480f


[SPARK-2464][Streaming] Fixed Twitter stream stopping bug

Stopping the Twitter Receiver would call twitter4j's TwitterStream.shutdown, 
which in turn causes an Exception to be thrown to the listener. This exception 
caused the Receiver to be restarted. This patch check whether the receiver was 
stopped or not, and accordingly restarts on exception.

Author: Tathagata Das tathagata.das1...@gmail.com

Closes #1577 from tdas/twitter-stop and squashes the following commits:

011b525 [Tathagata Das] Fixed Twitter stream stopping bug.


Project: http://git-wip-us.apache.org/repos/asf/spark/repo
Commit: http://git-wip-us.apache.org/repos/asf/spark/commit/a45d5480
Tree: http://git-wip-us.apache.org/repos/asf/spark/tree/a45d5480
Diff: http://git-wip-us.apache.org/repos/asf/spark/diff/a45d5480

Branch: refs/heads/master
Commit: a45d5480f65d2e969fc7fbd8f358b1717fb99bef
Parents: fec641b
Author: Tathagata Das tathagata.das1...@gmail.com
Authored: Thu Jul 24 15:59:09 2014 -0700
Committer: Tathagata Das tathagata.das1...@gmail.com
Committed: Thu Jul 24 15:59:09 2014 -0700

--
 .../spark/streaming/twitter/TwitterInputDStream.scala   | 9 +++--
 1 file changed, 7 insertions(+), 2 deletions(-)
--


http://git-wip-us.apache.org/repos/asf/spark/blob/a45d5480/external/twitter/src/main/scala/org/apache/spark/streaming/twitter/TwitterInputDStream.scala
--
diff --git 
a/external/twitter/src/main/scala/org/apache/spark/streaming/twitter/TwitterInputDStream.scala
 
b/external/twitter/src/main/scala/org/apache/spark/streaming/twitter/TwitterInputDStream.scala
index 5ea2e55..4eacc47 100644
--- 
a/external/twitter/src/main/scala/org/apache/spark/streaming/twitter/TwitterInputDStream.scala
+++ 
b/external/twitter/src/main/scala/org/apache/spark/streaming/twitter/TwitterInputDStream.scala
@@ -63,7 +63,8 @@ class TwitterReceiver(
 storageLevel: StorageLevel
   ) extends Receiver[Status](storageLevel) with Logging {
 
-  private var twitterStream: TwitterStream = _
+  @volatile private var twitterStream: TwitterStream = _
+  @volatile private var stopped = false
 
   def onStart() {
 try {
@@ -78,7 +79,9 @@ class TwitterReceiver(
 def onScrubGeo(l: Long, l1: Long) {}
 def onStallWarning(stallWarning: StallWarning) {}
 def onException(e: Exception) {
-  restart(Error receiving tweets, e)
+  if (!stopped) {
+restart(Error receiving tweets, e)
+  }
 }
   })
 
@@ -91,12 +94,14 @@ class TwitterReceiver(
   }
   setTwitterStream(newTwitterStream)
   logInfo(Twitter receiver started)
+  stopped = false
 } catch {
   case e: Exception = restart(Error starting Twitter stream, e)
 }
   }
 
   def onStop() {
+stopped = true
 setTwitterStream(null)
 logInfo(Twitter receiver stopped)
   }



git commit: [SPARK-2464][Streaming] Fixed Twitter stream stopping bug

2014-07-24 Thread tdas
Repository: spark
Updated Branches:
  refs/heads/branch-1.0 91241592a - 53b4e0f95


[SPARK-2464][Streaming] Fixed Twitter stream stopping bug

Stopping the Twitter Receiver would call twitter4j's TwitterStream.shutdown, 
which in turn causes an Exception to be thrown to the listener. This exception 
caused the Receiver to be restarted. This patch check whether the receiver was 
stopped or not, and accordingly restarts on exception.

Author: Tathagata Das tathagata.das1...@gmail.com

Closes #1577 from tdas/twitter-stop and squashes the following commits:

011b525 [Tathagata Das] Fixed Twitter stream stopping bug.

(cherry picked from commit a45d5480f65d2e969fc7fbd8f358b1717fb99bef)
Signed-off-by: Tathagata Das tathagata.das1...@gmail.com


Project: http://git-wip-us.apache.org/repos/asf/spark/repo
Commit: http://git-wip-us.apache.org/repos/asf/spark/commit/53b4e0f9
Tree: http://git-wip-us.apache.org/repos/asf/spark/tree/53b4e0f9
Diff: http://git-wip-us.apache.org/repos/asf/spark/diff/53b4e0f9

Branch: refs/heads/branch-1.0
Commit: 53b4e0f95750f371db14f8da4b5c4a1c38301710
Parents: 9124159
Author: Tathagata Das tathagata.das1...@gmail.com
Authored: Thu Jul 24 15:59:09 2014 -0700
Committer: Tathagata Das tathagata.das1...@gmail.com
Committed: Thu Jul 24 16:00:49 2014 -0700

--
 .../spark/streaming/twitter/TwitterInputDStream.scala   | 9 +++--
 1 file changed, 7 insertions(+), 2 deletions(-)
--


http://git-wip-us.apache.org/repos/asf/spark/blob/53b4e0f9/external/twitter/src/main/scala/org/apache/spark/streaming/twitter/TwitterInputDStream.scala
--
diff --git 
a/external/twitter/src/main/scala/org/apache/spark/streaming/twitter/TwitterInputDStream.scala
 
b/external/twitter/src/main/scala/org/apache/spark/streaming/twitter/TwitterInputDStream.scala
index 5ea2e55..4eacc47 100644
--- 
a/external/twitter/src/main/scala/org/apache/spark/streaming/twitter/TwitterInputDStream.scala
+++ 
b/external/twitter/src/main/scala/org/apache/spark/streaming/twitter/TwitterInputDStream.scala
@@ -63,7 +63,8 @@ class TwitterReceiver(
 storageLevel: StorageLevel
   ) extends Receiver[Status](storageLevel) with Logging {
 
-  private var twitterStream: TwitterStream = _
+  @volatile private var twitterStream: TwitterStream = _
+  @volatile private var stopped = false
 
   def onStart() {
 try {
@@ -78,7 +79,9 @@ class TwitterReceiver(
 def onScrubGeo(l: Long, l1: Long) {}
 def onStallWarning(stallWarning: StallWarning) {}
 def onException(e: Exception) {
-  restart(Error receiving tweets, e)
+  if (!stopped) {
+restart(Error receiving tweets, e)
+  }
 }
   })
 
@@ -91,12 +94,14 @@ class TwitterReceiver(
   }
   setTwitterStream(newTwitterStream)
   logInfo(Twitter receiver started)
+  stopped = false
 } catch {
   case e: Exception = restart(Error starting Twitter stream, e)
 }
   }
 
   def onStop() {
+stopped = true
 setTwitterStream(null)
 logInfo(Twitter receiver stopped)
   }



[3/3] git commit: Updating versions for 1.0.2 release.

2014-07-24 Thread tdas
Updating versions for 1.0.2 release.


Project: http://git-wip-us.apache.org/repos/asf/spark/repo
Commit: http://git-wip-us.apache.org/repos/asf/spark/commit/70109da2
Tree: http://git-wip-us.apache.org/repos/asf/spark/tree/70109da2
Diff: http://git-wip-us.apache.org/repos/asf/spark/diff/70109da2

Branch: refs/heads/branch-1.0
Commit: 70109da212343601d428252e9d298f6affa457f3
Parents: b1e1917
Author: Tathagata Das tathagata.das1...@gmail.com
Authored: Fri Jul 25 03:09:36 2014 +
Committer: Tathagata Das tathagata.das1...@gmail.com
Committed: Fri Jul 25 03:09:36 2014 +

--
 ec2/spark_ec2.py   | 2 +-
 project/SparkBuild.scala   | 2 +-
 project/project/SparkPluginBuild.scala | 2 +-
 python/pyspark/shell.py| 2 +-
 repl/src/main/scala/org/apache/spark/repl/SparkILoopInit.scala | 2 +-
 yarn/alpha/pom.xml | 2 +-
 6 files changed, 6 insertions(+), 6 deletions(-)
--


http://git-wip-us.apache.org/repos/asf/spark/blob/70109da2/ec2/spark_ec2.py
--
diff --git a/ec2/spark_ec2.py b/ec2/spark_ec2.py
index 83fe920..348277a 100755
--- a/ec2/spark_ec2.py
+++ b/ec2/spark_ec2.py
@@ -164,7 +164,7 @@ def is_active(instance):
 # Return correct versions of Spark and Shark, given the supplied Spark version
 def get_spark_shark_version(opts):
   spark_shark_map = {0.7.3: 0.7.1, 0.8.0: 0.8.0, 0.8.1: 0.8.1, 
0.9.0: 0.9.0, 
-0.9.1: 0.9.1, 1.0.0: 1.0.0, 1.0.1: 1.0.0}
+0.9.1: 0.9.1, 1.0.0: 1.0.0, 1.0.1: 1.0.0, 1.0.2: 1.0.0}
   version = opts.spark_version.replace(v, )
   if version not in spark_shark_map:
 print  stderr, Don't know about Spark version: %s % version

http://git-wip-us.apache.org/repos/asf/spark/blob/70109da2/project/SparkBuild.scala
--
diff --git a/project/SparkBuild.scala b/project/SparkBuild.scala
index e8edbf0..5b687e7 100644
--- a/project/SparkBuild.scala
+++ b/project/SparkBuild.scala
@@ -32,7 +32,7 @@ import scala.collection.JavaConversions._
 // import com.jsuereth.pgp.sbtplugin.PgpKeys._
 
 object SparkBuild extends Build {
-  val SPARK_VERSION = 1.0.2-SNAPSHOT
+  val SPARK_VERSION = 1.0.2
   val SPARK_VERSION_SHORT = SPARK_VERSION.replaceAll(-SNAPSHOT, )
 
   // Hadoop version to build against. For example, 1.0.4 for Apache 
releases, or

http://git-wip-us.apache.org/repos/asf/spark/blob/70109da2/project/project/SparkPluginBuild.scala
--
diff --git a/project/project/SparkPluginBuild.scala 
b/project/project/SparkPluginBuild.scala
index aca93e3..722d26f 100644
--- a/project/project/SparkPluginBuild.scala
+++ b/project/project/SparkPluginBuild.scala
@@ -26,7 +26,7 @@ import sbt.Keys._
 object SparkPluginDef extends Build {
   lazy val root = Project(plugins, file(.)) dependsOn(sparkStyle)
   lazy val sparkStyle = Project(spark-style, file(spark-style), settings = 
styleSettings)
-  val sparkVersion = 1.0.2-SNAPSHOT
+  val sparkVersion = 1.0.2
   // There is actually no need to publish this artifact.
   def styleSettings = Defaults.defaultSettings ++ Seq (
 name :=  spark-style,

http://git-wip-us.apache.org/repos/asf/spark/blob/70109da2/python/pyspark/shell.py
--
diff --git a/python/pyspark/shell.py b/python/pyspark/shell.py
index 5fa8178..e6c3306 100644
--- a/python/pyspark/shell.py
+++ b/python/pyspark/shell.py
@@ -46,7 +46,7 @@ print(Welcome to
     __
  / __/__  ___ _/ /__
 _\ \/ _ \/ _ `/ __/  '_/
-   /__ / .__/\_,_/_/ /_/\_\   version 1.0.2-SNAPSHOT
+   /__ / .__/\_,_/_/ /_/\_\   version 1.0.2
   /_/
 )
 print(Using Python version %s (%s, %s) % (

http://git-wip-us.apache.org/repos/asf/spark/blob/70109da2/repl/src/main/scala/org/apache/spark/repl/SparkILoopInit.scala
--
diff --git a/repl/src/main/scala/org/apache/spark/repl/SparkILoopInit.scala 
b/repl/src/main/scala/org/apache/spark/repl/SparkILoopInit.scala
index d7c05c5..9490195 100644
--- a/repl/src/main/scala/org/apache/spark/repl/SparkILoopInit.scala
+++ b/repl/src/main/scala/org/apache/spark/repl/SparkILoopInit.scala
@@ -26,7 +26,7 @@ trait SparkILoopInit {
     __
  / __/__  ___ _/ /__
 _\ \/ _ \/ _ `/ __/  '_/
-   /___/ .__/\_,_/_/ /_/\_\   version 1.0.2-SNAPSHOT
+   /___/ .__/\_,_/_/ /_/\_\   version 1.0.2
   /_/
 )
 import Properties._

http://git-wip-us.apache.org/repos/asf/spark/blob/70109da2/yarn/alpha/pom.xml
--

[1/3] git commit: Revert [maven-release-plugin] prepare for next development iteration

2014-07-24 Thread tdas
Repository: spark
Updated Branches:
  refs/heads/branch-1.0 53b4e0f95 - 70109da21


Revert [maven-release-plugin] prepare for next development iteration

This reverts commit baf92a0f2119867b1be540085ebe9f1a1c411ae8.


Project: http://git-wip-us.apache.org/repos/asf/spark/repo
Commit: http://git-wip-us.apache.org/repos/asf/spark/commit/d10455c6
Tree: http://git-wip-us.apache.org/repos/asf/spark/tree/d10455c6
Diff: http://git-wip-us.apache.org/repos/asf/spark/diff/d10455c6

Branch: refs/heads/branch-1.0
Commit: d10455c6f1f25da165e9a0a7f1b1be62586b99c2
Parents: 53b4e0f
Author: Ubuntu ubu...@ip-172-31-4-51.us-west-2.compute.internal
Authored: Fri Jul 25 02:20:14 2014 +
Committer: Ubuntu ubu...@ip-172-31-4-51.us-west-2.compute.internal
Committed: Fri Jul 25 02:20:14 2014 +

--
 assembly/pom.xml  | 2 +-
 bagel/pom.xml | 2 +-
 core/pom.xml  | 2 +-
 examples/pom.xml  | 2 +-
 external/flume/pom.xml| 2 +-
 external/kafka/pom.xml| 2 +-
 external/mqtt/pom.xml | 2 +-
 external/twitter/pom.xml  | 2 +-
 external/zeromq/pom.xml   | 2 +-
 extras/spark-ganglia-lgpl/pom.xml | 2 +-
 graphx/pom.xml| 2 +-
 mllib/pom.xml | 2 +-
 pom.xml   | 4 ++--
 repl/pom.xml  | 2 +-
 sql/catalyst/pom.xml  | 2 +-
 sql/core/pom.xml  | 2 +-
 sql/hive/pom.xml  | 2 +-
 streaming/pom.xml | 2 +-
 tools/pom.xml | 2 +-
 yarn/pom.xml  | 2 +-
 yarn/stable/pom.xml   | 2 +-
 21 files changed, 22 insertions(+), 22 deletions(-)
--


http://git-wip-us.apache.org/repos/asf/spark/blob/d10455c6/assembly/pom.xml
--
diff --git a/assembly/pom.xml b/assembly/pom.xml
index fbb993c..6735379 100644
--- a/assembly/pom.xml
+++ b/assembly/pom.xml
@@ -21,7 +21,7 @@
   parent
 groupIdorg.apache.spark/groupId
 artifactIdspark-parent/artifactId
-version1.0.3-SNAPSHOT/version
+version1.0.2/version
 relativePath../pom.xml/relativePath
   /parent
 

http://git-wip-us.apache.org/repos/asf/spark/blob/d10455c6/bagel/pom.xml
--
diff --git a/bagel/pom.xml b/bagel/pom.xml
index 0f7b0d7..8a38b43 100644
--- a/bagel/pom.xml
+++ b/bagel/pom.xml
@@ -21,7 +21,7 @@
   parent
 groupIdorg.apache.spark/groupId
 artifactIdspark-parent/artifactId
-version1.0.3-SNAPSHOT/version
+version1.0.2/version
 relativePath../pom.xml/relativePath
   /parent
 

http://git-wip-us.apache.org/repos/asf/spark/blob/d10455c6/core/pom.xml
--
diff --git a/core/pom.xml b/core/pom.xml
index 59110c2..23eea6e 100644
--- a/core/pom.xml
+++ b/core/pom.xml
@@ -21,7 +21,7 @@
   parent
 groupIdorg.apache.spark/groupId
 artifactIdspark-parent/artifactId
-version1.0.3-SNAPSHOT/version
+version1.0.2/version
 relativePath../pom.xml/relativePath
   /parent
 

http://git-wip-us.apache.org/repos/asf/spark/blob/d10455c6/examples/pom.xml
--
diff --git a/examples/pom.xml b/examples/pom.xml
index 08a9ddd..d158a75 100644
--- a/examples/pom.xml
+++ b/examples/pom.xml
@@ -21,7 +21,7 @@
   parent
 groupIdorg.apache.spark/groupId
 artifactIdspark-parent/artifactId
-version1.0.3-SNAPSHOT/version
+version1.0.2/version
 relativePath../pom.xml/relativePath
   /parent
 

http://git-wip-us.apache.org/repos/asf/spark/blob/d10455c6/external/flume/pom.xml
--
diff --git a/external/flume/pom.xml b/external/flume/pom.xml
index a486354..f5f0d54 100644
--- a/external/flume/pom.xml
+++ b/external/flume/pom.xml
@@ -21,7 +21,7 @@
   parent
 groupIdorg.apache.spark/groupId
 artifactIdspark-parent/artifactId
-version1.0.3-SNAPSHOT/version
+version1.0.2/version
 relativePath../../pom.xml/relativePath
   /parent
 

http://git-wip-us.apache.org/repos/asf/spark/blob/d10455c6/external/kafka/pom.xml
--
diff --git a/external/kafka/pom.xml b/external/kafka/pom.xml
index f68ad12..8bc5c03 100644
--- a/external/kafka/pom.xml
+++ b/external/kafka/pom.xml
@@ -21,7 +21,7 @@
   parent
 groupIdorg.apache.spark/groupId
 artifactIdspark-parent/artifactId
-version1.0.3-SNAPSHOT/version
+version1.0.2/version
 relativePath../../pom.xml/relativePath
   /parent
 

http://git-wip-us.apache.org/repos/asf/spark/blob/d10455c6/external/mqtt/pom.xml

git commit: [maven-release-plugin] prepare release v1.0.2-rc1

2014-07-25 Thread tdas
Repository: spark
Updated Branches:
  refs/heads/branch-1.0 797c663ae - 919c87f26


[maven-release-plugin] prepare release v1.0.2-rc1


Project: http://git-wip-us.apache.org/repos/asf/spark/repo
Commit: http://git-wip-us.apache.org/repos/asf/spark/commit/919c87f2
Tree: http://git-wip-us.apache.org/repos/asf/spark/tree/919c87f2
Diff: http://git-wip-us.apache.org/repos/asf/spark/diff/919c87f2

Branch: refs/heads/branch-1.0
Commit: 919c87f26a2655bfd5ae03958915b6804367c1d6
Parents: 797c663
Author: Tathagata Das tathagata.das1...@gmail.com
Authored: Fri Jul 25 11:30:01 2014 +
Committer: Tathagata Das tathagata.das1...@gmail.com
Committed: Fri Jul 25 11:30:01 2014 +

--
 assembly/pom.xml  | 2 +-
 bagel/pom.xml | 2 +-
 core/pom.xml  | 2 +-
 examples/pom.xml  | 2 +-
 external/flume/pom.xml| 2 +-
 external/kafka/pom.xml| 2 +-
 external/mqtt/pom.xml | 2 +-
 external/twitter/pom.xml  | 2 +-
 external/zeromq/pom.xml   | 2 +-
 extras/spark-ganglia-lgpl/pom.xml | 2 +-
 graphx/pom.xml| 2 +-
 mllib/pom.xml | 2 +-
 pom.xml   | 4 ++--
 repl/pom.xml  | 2 +-
 sql/catalyst/pom.xml  | 2 +-
 sql/core/pom.xml  | 2 +-
 sql/hive/pom.xml  | 2 +-
 streaming/pom.xml | 2 +-
 tools/pom.xml | 2 +-
 yarn/pom.xml  | 2 +-
 yarn/stable/pom.xml   | 2 +-
 21 files changed, 22 insertions(+), 22 deletions(-)
--


http://git-wip-us.apache.org/repos/asf/spark/blob/919c87f2/assembly/pom.xml
--
diff --git a/assembly/pom.xml b/assembly/pom.xml
index e290e79..6735379 100644
--- a/assembly/pom.xml
+++ b/assembly/pom.xml
@@ -21,7 +21,7 @@
   parent
 groupIdorg.apache.spark/groupId
 artifactIdspark-parent/artifactId
-version1.0.2-SNAPSHOT/version
+version1.0.2/version
 relativePath../pom.xml/relativePath
   /parent
 

http://git-wip-us.apache.org/repos/asf/spark/blob/919c87f2/bagel/pom.xml
--
diff --git a/bagel/pom.xml b/bagel/pom.xml
index c8ad40f..8a38b43 100644
--- a/bagel/pom.xml
+++ b/bagel/pom.xml
@@ -21,7 +21,7 @@
   parent
 groupIdorg.apache.spark/groupId
 artifactIdspark-parent/artifactId
-version1.0.2-SNAPSHOT/version
+version1.0.2/version
 relativePath../pom.xml/relativePath
   /parent
 

http://git-wip-us.apache.org/repos/asf/spark/blob/919c87f2/core/pom.xml
--
diff --git a/core/pom.xml b/core/pom.xml
index 2302d7b..23eea6e 100644
--- a/core/pom.xml
+++ b/core/pom.xml
@@ -21,7 +21,7 @@
   parent
 groupIdorg.apache.spark/groupId
 artifactIdspark-parent/artifactId
-version1.0.2-SNAPSHOT/version
+version1.0.2/version
 relativePath../pom.xml/relativePath
   /parent
 

http://git-wip-us.apache.org/repos/asf/spark/blob/919c87f2/examples/pom.xml
--
diff --git a/examples/pom.xml b/examples/pom.xml
index 9156a11..d158a75 100644
--- a/examples/pom.xml
+++ b/examples/pom.xml
@@ -21,7 +21,7 @@
   parent
 groupIdorg.apache.spark/groupId
 artifactIdspark-parent/artifactId
-version1.0.2-SNAPSHOT/version
+version1.0.2/version
 relativePath../pom.xml/relativePath
   /parent
 

http://git-wip-us.apache.org/repos/asf/spark/blob/919c87f2/external/flume/pom.xml
--
diff --git a/external/flume/pom.xml b/external/flume/pom.xml
index 1cefa15..f5f0d54 100644
--- a/external/flume/pom.xml
+++ b/external/flume/pom.xml
@@ -21,7 +21,7 @@
   parent
 groupIdorg.apache.spark/groupId
 artifactIdspark-parent/artifactId
-version1.0.2-SNAPSHOT/version
+version1.0.2/version
 relativePath../../pom.xml/relativePath
   /parent
 

http://git-wip-us.apache.org/repos/asf/spark/blob/919c87f2/external/kafka/pom.xml
--
diff --git a/external/kafka/pom.xml b/external/kafka/pom.xml
index cc05e69..8bc5c03 100644
--- a/external/kafka/pom.xml
+++ b/external/kafka/pom.xml
@@ -21,7 +21,7 @@
   parent
 groupIdorg.apache.spark/groupId
 artifactIdspark-parent/artifactId
-version1.0.2-SNAPSHOT/version
+version1.0.2/version
 relativePath../../pom.xml/relativePath
   /parent
 

http://git-wip-us.apache.org/repos/asf/spark/blob/919c87f2/external/mqtt/pom.xml
--
diff --git a/external/mqtt/pom.xml b/external/mqtt/pom.xml
index ab5afa2..ca022e0 100644
--- 

Git Push Summary

2014-07-25 Thread tdas
Repository: spark
Updated Tags:  refs/tags/v1.0.2-rc1 [created] 86a5fc071


git commit: [maven-release-plugin] prepare for next development iteration

2014-07-25 Thread tdas
Repository: spark
Updated Branches:
  refs/heads/branch-1.0 919c87f26 - edbd02fc6


[maven-release-plugin] prepare for next development iteration


Project: http://git-wip-us.apache.org/repos/asf/spark/repo
Commit: http://git-wip-us.apache.org/repos/asf/spark/commit/edbd02fc
Tree: http://git-wip-us.apache.org/repos/asf/spark/tree/edbd02fc
Diff: http://git-wip-us.apache.org/repos/asf/spark/diff/edbd02fc

Branch: refs/heads/branch-1.0
Commit: edbd02fc6873676e080101d407916efb64bdf71a
Parents: 919c87f
Author: Tathagata Das tathagata.das1...@gmail.com
Authored: Fri Jul 25 11:30:10 2014 +
Committer: Tathagata Das tathagata.das1...@gmail.com
Committed: Fri Jul 25 11:30:10 2014 +

--
 assembly/pom.xml  | 2 +-
 bagel/pom.xml | 2 +-
 core/pom.xml  | 2 +-
 examples/pom.xml  | 2 +-
 external/flume/pom.xml| 2 +-
 external/kafka/pom.xml| 2 +-
 external/mqtt/pom.xml | 2 +-
 external/twitter/pom.xml  | 2 +-
 external/zeromq/pom.xml   | 2 +-
 extras/spark-ganglia-lgpl/pom.xml | 2 +-
 graphx/pom.xml| 2 +-
 mllib/pom.xml | 2 +-
 pom.xml   | 4 ++--
 repl/pom.xml  | 2 +-
 sql/catalyst/pom.xml  | 2 +-
 sql/core/pom.xml  | 2 +-
 sql/hive/pom.xml  | 2 +-
 streaming/pom.xml | 2 +-
 tools/pom.xml | 2 +-
 yarn/pom.xml  | 2 +-
 yarn/stable/pom.xml   | 2 +-
 21 files changed, 22 insertions(+), 22 deletions(-)
--


http://git-wip-us.apache.org/repos/asf/spark/blob/edbd02fc/assembly/pom.xml
--
diff --git a/assembly/pom.xml b/assembly/pom.xml
index 6735379..fbb993c 100644
--- a/assembly/pom.xml
+++ b/assembly/pom.xml
@@ -21,7 +21,7 @@
   parent
 groupIdorg.apache.spark/groupId
 artifactIdspark-parent/artifactId
-version1.0.2/version
+version1.0.3-SNAPSHOT/version
 relativePath../pom.xml/relativePath
   /parent
 

http://git-wip-us.apache.org/repos/asf/spark/blob/edbd02fc/bagel/pom.xml
--
diff --git a/bagel/pom.xml b/bagel/pom.xml
index 8a38b43..0f7b0d7 100644
--- a/bagel/pom.xml
+++ b/bagel/pom.xml
@@ -21,7 +21,7 @@
   parent
 groupIdorg.apache.spark/groupId
 artifactIdspark-parent/artifactId
-version1.0.2/version
+version1.0.3-SNAPSHOT/version
 relativePath../pom.xml/relativePath
   /parent
 

http://git-wip-us.apache.org/repos/asf/spark/blob/edbd02fc/core/pom.xml
--
diff --git a/core/pom.xml b/core/pom.xml
index 23eea6e..59110c2 100644
--- a/core/pom.xml
+++ b/core/pom.xml
@@ -21,7 +21,7 @@
   parent
 groupIdorg.apache.spark/groupId
 artifactIdspark-parent/artifactId
-version1.0.2/version
+version1.0.3-SNAPSHOT/version
 relativePath../pom.xml/relativePath
   /parent
 

http://git-wip-us.apache.org/repos/asf/spark/blob/edbd02fc/examples/pom.xml
--
diff --git a/examples/pom.xml b/examples/pom.xml
index d158a75..08a9ddd 100644
--- a/examples/pom.xml
+++ b/examples/pom.xml
@@ -21,7 +21,7 @@
   parent
 groupIdorg.apache.spark/groupId
 artifactIdspark-parent/artifactId
-version1.0.2/version
+version1.0.3-SNAPSHOT/version
 relativePath../pom.xml/relativePath
   /parent
 

http://git-wip-us.apache.org/repos/asf/spark/blob/edbd02fc/external/flume/pom.xml
--
diff --git a/external/flume/pom.xml b/external/flume/pom.xml
index f5f0d54..a486354 100644
--- a/external/flume/pom.xml
+++ b/external/flume/pom.xml
@@ -21,7 +21,7 @@
   parent
 groupIdorg.apache.spark/groupId
 artifactIdspark-parent/artifactId
-version1.0.2/version
+version1.0.3-SNAPSHOT/version
 relativePath../../pom.xml/relativePath
   /parent
 

http://git-wip-us.apache.org/repos/asf/spark/blob/edbd02fc/external/kafka/pom.xml
--
diff --git a/external/kafka/pom.xml b/external/kafka/pom.xml
index 8bc5c03..f68ad12 100644
--- a/external/kafka/pom.xml
+++ b/external/kafka/pom.xml
@@ -21,7 +21,7 @@
   parent
 groupIdorg.apache.spark/groupId
 artifactIdspark-parent/artifactId
-version1.0.2/version
+version1.0.3-SNAPSHOT/version
 relativePath../../pom.xml/relativePath
   /parent
 

http://git-wip-us.apache.org/repos/asf/spark/blob/edbd02fc/external/mqtt/pom.xml
--
diff --git a/external/mqtt/pom.xml b/external/mqtt/pom.xml
index ca022e0..b047b0f 100644
--- 

[2/2] git commit: Revert [maven-release-plugin] prepare release v1.0.2-rc1

2014-07-25 Thread tdas
Revert [maven-release-plugin] prepare release v1.0.2-rc1

This reverts commit 919c87f26a2655bfd5ae03958915b6804367c1d6.


Project: http://git-wip-us.apache.org/repos/asf/spark/repo
Commit: http://git-wip-us.apache.org/repos/asf/spark/commit/01fc6d88
Tree: http://git-wip-us.apache.org/repos/asf/spark/tree/01fc6d88
Diff: http://git-wip-us.apache.org/repos/asf/spark/diff/01fc6d88

Branch: refs/heads/branch-1.0
Commit: 01fc6d885aab03e12d34b6823e5a63ce87013e26
Parents: d9ccf7f
Author: Tathagata Das tathagata.das1...@gmail.com
Authored: Fri Jul 25 18:18:06 2014 +
Committer: Tathagata Das tathagata.das1...@gmail.com
Committed: Fri Jul 25 18:18:06 2014 +

--
 assembly/pom.xml  | 2 +-
 bagel/pom.xml | 2 +-
 core/pom.xml  | 2 +-
 examples/pom.xml  | 2 +-
 external/flume/pom.xml| 2 +-
 external/kafka/pom.xml| 2 +-
 external/mqtt/pom.xml | 2 +-
 external/twitter/pom.xml  | 2 +-
 external/zeromq/pom.xml   | 2 +-
 extras/spark-ganglia-lgpl/pom.xml | 2 +-
 graphx/pom.xml| 2 +-
 mllib/pom.xml | 2 +-
 pom.xml   | 4 ++--
 repl/pom.xml  | 2 +-
 sql/catalyst/pom.xml  | 2 +-
 sql/core/pom.xml  | 2 +-
 sql/hive/pom.xml  | 2 +-
 streaming/pom.xml | 2 +-
 tools/pom.xml | 2 +-
 yarn/pom.xml  | 2 +-
 yarn/stable/pom.xml   | 2 +-
 21 files changed, 22 insertions(+), 22 deletions(-)
--


http://git-wip-us.apache.org/repos/asf/spark/blob/01fc6d88/assembly/pom.xml
--
diff --git a/assembly/pom.xml b/assembly/pom.xml
index 6735379..e290e79 100644
--- a/assembly/pom.xml
+++ b/assembly/pom.xml
@@ -21,7 +21,7 @@
   parent
 groupIdorg.apache.spark/groupId
 artifactIdspark-parent/artifactId
-version1.0.2/version
+version1.0.2-SNAPSHOT/version
 relativePath../pom.xml/relativePath
   /parent
 

http://git-wip-us.apache.org/repos/asf/spark/blob/01fc6d88/bagel/pom.xml
--
diff --git a/bagel/pom.xml b/bagel/pom.xml
index 8a38b43..c8ad40f 100644
--- a/bagel/pom.xml
+++ b/bagel/pom.xml
@@ -21,7 +21,7 @@
   parent
 groupIdorg.apache.spark/groupId
 artifactIdspark-parent/artifactId
-version1.0.2/version
+version1.0.2-SNAPSHOT/version
 relativePath../pom.xml/relativePath
   /parent
 

http://git-wip-us.apache.org/repos/asf/spark/blob/01fc6d88/core/pom.xml
--
diff --git a/core/pom.xml b/core/pom.xml
index 23eea6e..2302d7b 100644
--- a/core/pom.xml
+++ b/core/pom.xml
@@ -21,7 +21,7 @@
   parent
 groupIdorg.apache.spark/groupId
 artifactIdspark-parent/artifactId
-version1.0.2/version
+version1.0.2-SNAPSHOT/version
 relativePath../pom.xml/relativePath
   /parent
 

http://git-wip-us.apache.org/repos/asf/spark/blob/01fc6d88/examples/pom.xml
--
diff --git a/examples/pom.xml b/examples/pom.xml
index d158a75..9156a11 100644
--- a/examples/pom.xml
+++ b/examples/pom.xml
@@ -21,7 +21,7 @@
   parent
 groupIdorg.apache.spark/groupId
 artifactIdspark-parent/artifactId
-version1.0.2/version
+version1.0.2-SNAPSHOT/version
 relativePath../pom.xml/relativePath
   /parent
 

http://git-wip-us.apache.org/repos/asf/spark/blob/01fc6d88/external/flume/pom.xml
--
diff --git a/external/flume/pom.xml b/external/flume/pom.xml
index f5f0d54..1cefa15 100644
--- a/external/flume/pom.xml
+++ b/external/flume/pom.xml
@@ -21,7 +21,7 @@
   parent
 groupIdorg.apache.spark/groupId
 artifactIdspark-parent/artifactId
-version1.0.2/version
+version1.0.2-SNAPSHOT/version
 relativePath../../pom.xml/relativePath
   /parent
 

http://git-wip-us.apache.org/repos/asf/spark/blob/01fc6d88/external/kafka/pom.xml
--
diff --git a/external/kafka/pom.xml b/external/kafka/pom.xml
index 8bc5c03..cc05e69 100644
--- a/external/kafka/pom.xml
+++ b/external/kafka/pom.xml
@@ -21,7 +21,7 @@
   parent
 groupIdorg.apache.spark/groupId
 artifactIdspark-parent/artifactId
-version1.0.2/version
+version1.0.2-SNAPSHOT/version
 relativePath../../pom.xml/relativePath
   /parent
 

http://git-wip-us.apache.org/repos/asf/spark/blob/01fc6d88/external/mqtt/pom.xml
--
diff --git a/external/mqtt/pom.xml b/external/mqtt/pom.xml
index ca022e0..ab5afa2 100644
--- a/external/mqtt/pom.xml

git commit: [maven-release-plugin] prepare release v1.0.2-rc1

2014-07-25 Thread tdas
Repository: spark
Updated Branches:
  refs/heads/branch-1.0 01fc6d885 - 08f601328


[maven-release-plugin] prepare release v1.0.2-rc1


Project: http://git-wip-us.apache.org/repos/asf/spark/repo
Commit: http://git-wip-us.apache.org/repos/asf/spark/commit/08f60132
Tree: http://git-wip-us.apache.org/repos/asf/spark/tree/08f60132
Diff: http://git-wip-us.apache.org/repos/asf/spark/diff/08f60132

Branch: refs/heads/branch-1.0
Commit: 08f601328ad9e7334ef7deb3a9fff1343a3c4f30
Parents: 01fc6d8
Author: Tathagata Das tathagata.das1...@gmail.com
Authored: Fri Jul 25 18:43:18 2014 +
Committer: Tathagata Das tathagata.das1...@gmail.com
Committed: Fri Jul 25 18:43:18 2014 +

--
 assembly/pom.xml  | 2 +-
 bagel/pom.xml | 2 +-
 core/pom.xml  | 2 +-
 examples/pom.xml  | 2 +-
 external/flume/pom.xml| 2 +-
 external/kafka/pom.xml| 2 +-
 external/mqtt/pom.xml | 2 +-
 external/twitter/pom.xml  | 2 +-
 external/zeromq/pom.xml   | 2 +-
 extras/spark-ganglia-lgpl/pom.xml | 2 +-
 graphx/pom.xml| 2 +-
 mllib/pom.xml | 2 +-
 pom.xml   | 4 ++--
 repl/pom.xml  | 2 +-
 sql/catalyst/pom.xml  | 2 +-
 sql/core/pom.xml  | 2 +-
 sql/hive/pom.xml  | 2 +-
 streaming/pom.xml | 2 +-
 tools/pom.xml | 2 +-
 yarn/pom.xml  | 2 +-
 yarn/stable/pom.xml   | 2 +-
 21 files changed, 22 insertions(+), 22 deletions(-)
--


http://git-wip-us.apache.org/repos/asf/spark/blob/08f60132/assembly/pom.xml
--
diff --git a/assembly/pom.xml b/assembly/pom.xml
index e290e79..6735379 100644
--- a/assembly/pom.xml
+++ b/assembly/pom.xml
@@ -21,7 +21,7 @@
   parent
 groupIdorg.apache.spark/groupId
 artifactIdspark-parent/artifactId
-version1.0.2-SNAPSHOT/version
+version1.0.2/version
 relativePath../pom.xml/relativePath
   /parent
 

http://git-wip-us.apache.org/repos/asf/spark/blob/08f60132/bagel/pom.xml
--
diff --git a/bagel/pom.xml b/bagel/pom.xml
index c8ad40f..8a38b43 100644
--- a/bagel/pom.xml
+++ b/bagel/pom.xml
@@ -21,7 +21,7 @@
   parent
 groupIdorg.apache.spark/groupId
 artifactIdspark-parent/artifactId
-version1.0.2-SNAPSHOT/version
+version1.0.2/version
 relativePath../pom.xml/relativePath
   /parent
 

http://git-wip-us.apache.org/repos/asf/spark/blob/08f60132/core/pom.xml
--
diff --git a/core/pom.xml b/core/pom.xml
index 2302d7b..23eea6e 100644
--- a/core/pom.xml
+++ b/core/pom.xml
@@ -21,7 +21,7 @@
   parent
 groupIdorg.apache.spark/groupId
 artifactIdspark-parent/artifactId
-version1.0.2-SNAPSHOT/version
+version1.0.2/version
 relativePath../pom.xml/relativePath
   /parent
 

http://git-wip-us.apache.org/repos/asf/spark/blob/08f60132/examples/pom.xml
--
diff --git a/examples/pom.xml b/examples/pom.xml
index 9156a11..d158a75 100644
--- a/examples/pom.xml
+++ b/examples/pom.xml
@@ -21,7 +21,7 @@
   parent
 groupIdorg.apache.spark/groupId
 artifactIdspark-parent/artifactId
-version1.0.2-SNAPSHOT/version
+version1.0.2/version
 relativePath../pom.xml/relativePath
   /parent
 

http://git-wip-us.apache.org/repos/asf/spark/blob/08f60132/external/flume/pom.xml
--
diff --git a/external/flume/pom.xml b/external/flume/pom.xml
index 1cefa15..f5f0d54 100644
--- a/external/flume/pom.xml
+++ b/external/flume/pom.xml
@@ -21,7 +21,7 @@
   parent
 groupIdorg.apache.spark/groupId
 artifactIdspark-parent/artifactId
-version1.0.2-SNAPSHOT/version
+version1.0.2/version
 relativePath../../pom.xml/relativePath
   /parent
 

http://git-wip-us.apache.org/repos/asf/spark/blob/08f60132/external/kafka/pom.xml
--
diff --git a/external/kafka/pom.xml b/external/kafka/pom.xml
index cc05e69..8bc5c03 100644
--- a/external/kafka/pom.xml
+++ b/external/kafka/pom.xml
@@ -21,7 +21,7 @@
   parent
 groupIdorg.apache.spark/groupId
 artifactIdspark-parent/artifactId
-version1.0.2-SNAPSHOT/version
+version1.0.2/version
 relativePath../../pom.xml/relativePath
   /parent
 

http://git-wip-us.apache.org/repos/asf/spark/blob/08f60132/external/mqtt/pom.xml
--
diff --git a/external/mqtt/pom.xml b/external/mqtt/pom.xml
index ab5afa2..ca022e0 100644
--- 

git commit: [maven-release-plugin] prepare for next development iteration

2014-07-25 Thread tdas
Repository: spark
Updated Branches:
  refs/heads/branch-1.0 08f601328 - 54df1b8c3


[maven-release-plugin] prepare for next development iteration


Project: http://git-wip-us.apache.org/repos/asf/spark/repo
Commit: http://git-wip-us.apache.org/repos/asf/spark/commit/54df1b8c
Tree: http://git-wip-us.apache.org/repos/asf/spark/tree/54df1b8c
Diff: http://git-wip-us.apache.org/repos/asf/spark/diff/54df1b8c

Branch: refs/heads/branch-1.0
Commit: 54df1b8c31fa2de5b04ee4a5563706b2664f34f3
Parents: 08f6013
Author: Tathagata Das tathagata.das1...@gmail.com
Authored: Fri Jul 25 18:43:25 2014 +
Committer: Tathagata Das tathagata.das1...@gmail.com
Committed: Fri Jul 25 18:43:25 2014 +

--
 assembly/pom.xml  | 2 +-
 bagel/pom.xml | 2 +-
 core/pom.xml  | 2 +-
 examples/pom.xml  | 2 +-
 external/flume/pom.xml| 2 +-
 external/kafka/pom.xml| 2 +-
 external/mqtt/pom.xml | 2 +-
 external/twitter/pom.xml  | 2 +-
 external/zeromq/pom.xml   | 2 +-
 extras/spark-ganglia-lgpl/pom.xml | 2 +-
 graphx/pom.xml| 2 +-
 mllib/pom.xml | 2 +-
 pom.xml   | 4 ++--
 repl/pom.xml  | 2 +-
 sql/catalyst/pom.xml  | 2 +-
 sql/core/pom.xml  | 2 +-
 sql/hive/pom.xml  | 2 +-
 streaming/pom.xml | 2 +-
 tools/pom.xml | 2 +-
 yarn/pom.xml  | 2 +-
 yarn/stable/pom.xml   | 2 +-
 21 files changed, 22 insertions(+), 22 deletions(-)
--


http://git-wip-us.apache.org/repos/asf/spark/blob/54df1b8c/assembly/pom.xml
--
diff --git a/assembly/pom.xml b/assembly/pom.xml
index 6735379..fbb993c 100644
--- a/assembly/pom.xml
+++ b/assembly/pom.xml
@@ -21,7 +21,7 @@
   parent
 groupIdorg.apache.spark/groupId
 artifactIdspark-parent/artifactId
-version1.0.2/version
+version1.0.3-SNAPSHOT/version
 relativePath../pom.xml/relativePath
   /parent
 

http://git-wip-us.apache.org/repos/asf/spark/blob/54df1b8c/bagel/pom.xml
--
diff --git a/bagel/pom.xml b/bagel/pom.xml
index 8a38b43..0f7b0d7 100644
--- a/bagel/pom.xml
+++ b/bagel/pom.xml
@@ -21,7 +21,7 @@
   parent
 groupIdorg.apache.spark/groupId
 artifactIdspark-parent/artifactId
-version1.0.2/version
+version1.0.3-SNAPSHOT/version
 relativePath../pom.xml/relativePath
   /parent
 

http://git-wip-us.apache.org/repos/asf/spark/blob/54df1b8c/core/pom.xml
--
diff --git a/core/pom.xml b/core/pom.xml
index 23eea6e..59110c2 100644
--- a/core/pom.xml
+++ b/core/pom.xml
@@ -21,7 +21,7 @@
   parent
 groupIdorg.apache.spark/groupId
 artifactIdspark-parent/artifactId
-version1.0.2/version
+version1.0.3-SNAPSHOT/version
 relativePath../pom.xml/relativePath
   /parent
 

http://git-wip-us.apache.org/repos/asf/spark/blob/54df1b8c/examples/pom.xml
--
diff --git a/examples/pom.xml b/examples/pom.xml
index d158a75..08a9ddd 100644
--- a/examples/pom.xml
+++ b/examples/pom.xml
@@ -21,7 +21,7 @@
   parent
 groupIdorg.apache.spark/groupId
 artifactIdspark-parent/artifactId
-version1.0.2/version
+version1.0.3-SNAPSHOT/version
 relativePath../pom.xml/relativePath
   /parent
 

http://git-wip-us.apache.org/repos/asf/spark/blob/54df1b8c/external/flume/pom.xml
--
diff --git a/external/flume/pom.xml b/external/flume/pom.xml
index f5f0d54..a486354 100644
--- a/external/flume/pom.xml
+++ b/external/flume/pom.xml
@@ -21,7 +21,7 @@
   parent
 groupIdorg.apache.spark/groupId
 artifactIdspark-parent/artifactId
-version1.0.2/version
+version1.0.3-SNAPSHOT/version
 relativePath../../pom.xml/relativePath
   /parent
 

http://git-wip-us.apache.org/repos/asf/spark/blob/54df1b8c/external/kafka/pom.xml
--
diff --git a/external/kafka/pom.xml b/external/kafka/pom.xml
index 8bc5c03..f68ad12 100644
--- a/external/kafka/pom.xml
+++ b/external/kafka/pom.xml
@@ -21,7 +21,7 @@
   parent
 groupIdorg.apache.spark/groupId
 artifactIdspark-parent/artifactId
-version1.0.2/version
+version1.0.3-SNAPSHOT/version
 relativePath../../pom.xml/relativePath
   /parent
 

http://git-wip-us.apache.org/repos/asf/spark/blob/54df1b8c/external/mqtt/pom.xml
--
diff --git a/external/mqtt/pom.xml b/external/mqtt/pom.xml
index ca022e0..b047b0f 100644
--- 

Git Push Summary

2014-07-25 Thread tdas
Repository: spark
Updated Tags:  refs/tags/v1.0.2-rc1 [created] 6053ff1be


git commit: Updated CHANGES.txt

2014-07-25 Thread tdas
Repository: spark
Updated Branches:
  refs/heads/branch-1.0 54df1b8c3 - 76117baa7


Updated CHANGES.txt


Project: http://git-wip-us.apache.org/repos/asf/spark/repo
Commit: http://git-wip-us.apache.org/repos/asf/spark/commit/76117baa
Tree: http://git-wip-us.apache.org/repos/asf/spark/tree/76117baa
Diff: http://git-wip-us.apache.org/repos/asf/spark/diff/76117baa

Branch: refs/heads/branch-1.0
Commit: 76117baa7690deeaa765f672795be82ca51a891f
Parents: 54df1b8
Author: Tathagata Das tathagata.das1...@gmail.com
Authored: Fri Jul 25 20:50:50 2014 +
Committer: Tathagata Das tathagata.das1...@gmail.com
Committed: Fri Jul 25 20:50:50 2014 +

--
 CHANGES.txt   | 298 +
 dev/create-release/generate-changelist.py |   7 +-
 2 files changed, 301 insertions(+), 4 deletions(-)
--


http://git-wip-us.apache.org/repos/asf/spark/blob/76117baa/CHANGES.txt
--
diff --git a/CHANGES.txt b/CHANGES.txt
index a4ec2f1..18a31e0 100644
--- a/CHANGES.txt
+++ b/CHANGES.txt
@@ -1,6 +1,304 @@
 Spark Change Log
 
 
+Release 1.0.2
+
+  [SPARK-2529] Clean closures in foreach and foreachPartition.
+  Reynold Xin r...@apache.org
+  2014-07-25 01:10:05 -0700
+  Commit: 797c663, github.com/apache/spark/pull/1583
+
+  Updating versions for 1.0.2 release.
+  Tathagata Das tathagata.das1...@gmail.com
+  2014-07-25 03:09:36 +
+  Commit: 70109da
+
+  [SPARK-2464][Streaming] Fixed Twitter stream stopping bug
+  Tathagata Das tathagata.das1...@gmail.com
+  2014-07-24 15:59:09 -0700
+  Commit: 53b4e0f, github.com/apache/spark/pull/1577
+
+  [SPARK-2603][SQL] Remove unnecessary toMap and toList in converting Java 
collections to Scala collections JsonRDD.scala
+  Yin Huai h...@cse.ohio-state.edu
+  2014-07-24 11:19:19 -0700
+  Commit: 9124159, github.com/apache/spark/pull/1504
+
+  [SPARK-2658][SQL] Add rule for true = 1.
+  Michael Armbrust mich...@databricks.com
+  2014-07-23 22:52:49 -0700
+  Commit: 6b08046, github.com/apache/spark/pull/1556
+
+  [SPARK-2615] [SQL] Add Equal Sign == Support for HiveQl
+  Cheng Hao hao.ch...@intel.com
+  2014-07-22 18:13:28 -0700
+  Commit: c6421b6, github.com/apache/spark/pull/1522
+
+  [SPARK-2561][SQL] Fix apply schema
+  Michael Armbrust mich...@databricks.com
+  2014-07-21 18:18:17 -0700
+  Commit: 84bbfbd, github.com/apache/spark/pull/1470
+
+  [SPARK-2494] [PySpark] make hash of None consistant cross machines
+  Davies Liu davies@gmail.com
+  2014-07-21 11:59:54 -0700
+  Commit: cdcd467, github.com/apache/spark/pull/1371
+
+  Revert [SPARK-1199][REPL] Remove VALId and use the original import style 
for defined classes.
+  Patrick Wendell pwend...@gmail.com
+  2014-07-21 11:53:57 -0700
+  Commit: e0cc384
+
+  [SPARK-2598] RangePartitioner's binary search does not use the given Ordering
+  Reynold Xin r...@apache.org
+  2014-07-20 11:06:06 -0700
+  Commit: 480669f, github.com/apache/spark/pull/1500
+
+  [SPARK-2524] missing document about spark.deploy.retainedDrivers
+  lianhuiwang lianhuiwan...@gmail.com, Wang Lianhui 
lianhuiwan...@gmail.com, unknown administra...@taguswang-pc1.tencent.com
+  2014-07-19 20:46:59 -0700
+  Commit: 11670bf, github.com/apache/spark/pull/1443
+
+  Typo fix to the programming guide in the docs
+  Cesar Arevalo ce...@zephyrhealthinc.com
+  2014-07-19 20:20:07 -0700
+  Commit: a0624e8, github.com/apache/spark/pull/1495
+
+  [SPARK-2540] [SQL] Add HiveDecimal  HiveVarchar support in unwrapping data
+  Cheng Hao hao.ch...@intel.com
+  2014-07-18 16:38:11 -0500
+  Commit: 7611840, github.com/apache/spark/pull/1436
+
+  Added t2 instance types
+  Basit Mustafa basitmustafa@computes-things-for-basit.local
+  2014-07-18 12:23:47 -0700
+  Commit: 284bf10, github.com/apache/spark/pull/1446
+
+  [SPARK-2570] [SQL] Fix the bug of ClassCastException
+  Cheng Hao hao.ch...@intel.com
+  2014-07-17 23:25:01 -0700
+  Commit: d35837a, github.com/apache/spark/pull/1475
+
+  [SPARK-2534] Avoid pulling in the entire RDD in various operators 
(branch-1.0 backport)
+  Reynold Xin r...@apache.org
+  2014-07-17 16:33:30 -0700
+  Commit: 26c428a, github.com/apache/spark/pull/1469
+
+  [SPARK-2412] CoalescedRDD throws exception with certain pref locs
+  Aaron Davidson aa...@databricks.com
+  2014-07-17 01:01:14 -0700
+  Commit: 3bb5d2f, github.com/apache/spark/pull/1337
+
+  [SPARK-2154] Schedule next Driver when one completes (standalone mode)
+  Aaron Davidson aa...@databricks.com
+  2014-07-16 14:16:48 -0700
+  Commit: 0b0b895, github.com/apache/spark/pull/1405
+
+  SPARK-1097: Do not introduce deadlock while fixing concurrency bug
+  Aaron Davidson aa...@databricks.com
+  2014-07-16 14:10:17 -0700
+  Commit: 91e7a71, github.com/apache/spark/pull/1409
+
+  [SPARK-2518][SQL] Fix foldability of Substring expression.
+  Takuya 

[1/2] git commit: Revert [maven-release-plugin] prepare for next development iteration

2014-07-25 Thread tdas
Repository: spark
Updated Branches:
  refs/heads/branch-1.0 76117baa7 - 4be1dbd82


Revert [maven-release-plugin] prepare for next development iteration

This reverts commit 54df1b8c31fa2de5b04ee4a5563706b2664f34f3.


Project: http://git-wip-us.apache.org/repos/asf/spark/repo
Commit: http://git-wip-us.apache.org/repos/asf/spark/commit/57b5531a
Tree: http://git-wip-us.apache.org/repos/asf/spark/tree/57b5531a
Diff: http://git-wip-us.apache.org/repos/asf/spark/diff/57b5531a

Branch: refs/heads/branch-1.0
Commit: 57b5531a44ee0d0bd34007f2b5fc22dd733cb25a
Parents: 76117ba
Author: Tathagata Das tathagata.das1...@gmail.com
Authored: Fri Jul 25 20:54:02 2014 +
Committer: Tathagata Das tathagata.das1...@gmail.com
Committed: Fri Jul 25 20:54:02 2014 +

--
 assembly/pom.xml  | 2 +-
 bagel/pom.xml | 2 +-
 core/pom.xml  | 2 +-
 examples/pom.xml  | 2 +-
 external/flume/pom.xml| 2 +-
 external/kafka/pom.xml| 2 +-
 external/mqtt/pom.xml | 2 +-
 external/twitter/pom.xml  | 2 +-
 external/zeromq/pom.xml   | 2 +-
 extras/spark-ganglia-lgpl/pom.xml | 2 +-
 graphx/pom.xml| 2 +-
 mllib/pom.xml | 2 +-
 pom.xml   | 4 ++--
 repl/pom.xml  | 2 +-
 sql/catalyst/pom.xml  | 2 +-
 sql/core/pom.xml  | 2 +-
 sql/hive/pom.xml  | 2 +-
 streaming/pom.xml | 2 +-
 tools/pom.xml | 2 +-
 yarn/pom.xml  | 2 +-
 yarn/stable/pom.xml   | 2 +-
 21 files changed, 22 insertions(+), 22 deletions(-)
--


http://git-wip-us.apache.org/repos/asf/spark/blob/57b5531a/assembly/pom.xml
--
diff --git a/assembly/pom.xml b/assembly/pom.xml
index fbb993c..6735379 100644
--- a/assembly/pom.xml
+++ b/assembly/pom.xml
@@ -21,7 +21,7 @@
   parent
 groupIdorg.apache.spark/groupId
 artifactIdspark-parent/artifactId
-version1.0.3-SNAPSHOT/version
+version1.0.2/version
 relativePath../pom.xml/relativePath
   /parent
 

http://git-wip-us.apache.org/repos/asf/spark/blob/57b5531a/bagel/pom.xml
--
diff --git a/bagel/pom.xml b/bagel/pom.xml
index 0f7b0d7..8a38b43 100644
--- a/bagel/pom.xml
+++ b/bagel/pom.xml
@@ -21,7 +21,7 @@
   parent
 groupIdorg.apache.spark/groupId
 artifactIdspark-parent/artifactId
-version1.0.3-SNAPSHOT/version
+version1.0.2/version
 relativePath../pom.xml/relativePath
   /parent
 

http://git-wip-us.apache.org/repos/asf/spark/blob/57b5531a/core/pom.xml
--
diff --git a/core/pom.xml b/core/pom.xml
index 59110c2..23eea6e 100644
--- a/core/pom.xml
+++ b/core/pom.xml
@@ -21,7 +21,7 @@
   parent
 groupIdorg.apache.spark/groupId
 artifactIdspark-parent/artifactId
-version1.0.3-SNAPSHOT/version
+version1.0.2/version
 relativePath../pom.xml/relativePath
   /parent
 

http://git-wip-us.apache.org/repos/asf/spark/blob/57b5531a/examples/pom.xml
--
diff --git a/examples/pom.xml b/examples/pom.xml
index 08a9ddd..d158a75 100644
--- a/examples/pom.xml
+++ b/examples/pom.xml
@@ -21,7 +21,7 @@
   parent
 groupIdorg.apache.spark/groupId
 artifactIdspark-parent/artifactId
-version1.0.3-SNAPSHOT/version
+version1.0.2/version
 relativePath../pom.xml/relativePath
   /parent
 

http://git-wip-us.apache.org/repos/asf/spark/blob/57b5531a/external/flume/pom.xml
--
diff --git a/external/flume/pom.xml b/external/flume/pom.xml
index a486354..f5f0d54 100644
--- a/external/flume/pom.xml
+++ b/external/flume/pom.xml
@@ -21,7 +21,7 @@
   parent
 groupIdorg.apache.spark/groupId
 artifactIdspark-parent/artifactId
-version1.0.3-SNAPSHOT/version
+version1.0.2/version
 relativePath../../pom.xml/relativePath
   /parent
 

http://git-wip-us.apache.org/repos/asf/spark/blob/57b5531a/external/kafka/pom.xml
--
diff --git a/external/kafka/pom.xml b/external/kafka/pom.xml
index f68ad12..8bc5c03 100644
--- a/external/kafka/pom.xml
+++ b/external/kafka/pom.xml
@@ -21,7 +21,7 @@
   parent
 groupIdorg.apache.spark/groupId
 artifactIdspark-parent/artifactId
-version1.0.3-SNAPSHOT/version
+version1.0.2/version
 relativePath../../pom.xml/relativePath
   /parent
 

http://git-wip-us.apache.org/repos/asf/spark/blob/57b5531a/external/mqtt/pom.xml
--
diff --git 

[2/2] git commit: Revert [maven-release-plugin] prepare release v1.0.2-rc1

2014-07-25 Thread tdas
Revert [maven-release-plugin] prepare release v1.0.2-rc1

This reverts commit 08f601328ad9e7334ef7deb3a9fff1343a3c4f30.


Project: http://git-wip-us.apache.org/repos/asf/spark/repo
Commit: http://git-wip-us.apache.org/repos/asf/spark/commit/4be1dbd8
Tree: http://git-wip-us.apache.org/repos/asf/spark/tree/4be1dbd8
Diff: http://git-wip-us.apache.org/repos/asf/spark/diff/4be1dbd8

Branch: refs/heads/branch-1.0
Commit: 4be1dbd82302c0deb40f4f12f58129cc02a51842
Parents: 57b5531
Author: Tathagata Das tathagata.das1...@gmail.com
Authored: Fri Jul 25 20:54:14 2014 +
Committer: Tathagata Das tathagata.das1...@gmail.com
Committed: Fri Jul 25 20:54:14 2014 +

--
 assembly/pom.xml  | 2 +-
 bagel/pom.xml | 2 +-
 core/pom.xml  | 2 +-
 examples/pom.xml  | 2 +-
 external/flume/pom.xml| 2 +-
 external/kafka/pom.xml| 2 +-
 external/mqtt/pom.xml | 2 +-
 external/twitter/pom.xml  | 2 +-
 external/zeromq/pom.xml   | 2 +-
 extras/spark-ganglia-lgpl/pom.xml | 2 +-
 graphx/pom.xml| 2 +-
 mllib/pom.xml | 2 +-
 pom.xml   | 4 ++--
 repl/pom.xml  | 2 +-
 sql/catalyst/pom.xml  | 2 +-
 sql/core/pom.xml  | 2 +-
 sql/hive/pom.xml  | 2 +-
 streaming/pom.xml | 2 +-
 tools/pom.xml | 2 +-
 yarn/pom.xml  | 2 +-
 yarn/stable/pom.xml   | 2 +-
 21 files changed, 22 insertions(+), 22 deletions(-)
--


http://git-wip-us.apache.org/repos/asf/spark/blob/4be1dbd8/assembly/pom.xml
--
diff --git a/assembly/pom.xml b/assembly/pom.xml
index 6735379..e290e79 100644
--- a/assembly/pom.xml
+++ b/assembly/pom.xml
@@ -21,7 +21,7 @@
   parent
 groupIdorg.apache.spark/groupId
 artifactIdspark-parent/artifactId
-version1.0.2/version
+version1.0.2-SNAPSHOT/version
 relativePath../pom.xml/relativePath
   /parent
 

http://git-wip-us.apache.org/repos/asf/spark/blob/4be1dbd8/bagel/pom.xml
--
diff --git a/bagel/pom.xml b/bagel/pom.xml
index 8a38b43..c8ad40f 100644
--- a/bagel/pom.xml
+++ b/bagel/pom.xml
@@ -21,7 +21,7 @@
   parent
 groupIdorg.apache.spark/groupId
 artifactIdspark-parent/artifactId
-version1.0.2/version
+version1.0.2-SNAPSHOT/version
 relativePath../pom.xml/relativePath
   /parent
 

http://git-wip-us.apache.org/repos/asf/spark/blob/4be1dbd8/core/pom.xml
--
diff --git a/core/pom.xml b/core/pom.xml
index 23eea6e..2302d7b 100644
--- a/core/pom.xml
+++ b/core/pom.xml
@@ -21,7 +21,7 @@
   parent
 groupIdorg.apache.spark/groupId
 artifactIdspark-parent/artifactId
-version1.0.2/version
+version1.0.2-SNAPSHOT/version
 relativePath../pom.xml/relativePath
   /parent
 

http://git-wip-us.apache.org/repos/asf/spark/blob/4be1dbd8/examples/pom.xml
--
diff --git a/examples/pom.xml b/examples/pom.xml
index d158a75..9156a11 100644
--- a/examples/pom.xml
+++ b/examples/pom.xml
@@ -21,7 +21,7 @@
   parent
 groupIdorg.apache.spark/groupId
 artifactIdspark-parent/artifactId
-version1.0.2/version
+version1.0.2-SNAPSHOT/version
 relativePath../pom.xml/relativePath
   /parent
 

http://git-wip-us.apache.org/repos/asf/spark/blob/4be1dbd8/external/flume/pom.xml
--
diff --git a/external/flume/pom.xml b/external/flume/pom.xml
index f5f0d54..1cefa15 100644
--- a/external/flume/pom.xml
+++ b/external/flume/pom.xml
@@ -21,7 +21,7 @@
   parent
 groupIdorg.apache.spark/groupId
 artifactIdspark-parent/artifactId
-version1.0.2/version
+version1.0.2-SNAPSHOT/version
 relativePath../../pom.xml/relativePath
   /parent
 

http://git-wip-us.apache.org/repos/asf/spark/blob/4be1dbd8/external/kafka/pom.xml
--
diff --git a/external/kafka/pom.xml b/external/kafka/pom.xml
index 8bc5c03..cc05e69 100644
--- a/external/kafka/pom.xml
+++ b/external/kafka/pom.xml
@@ -21,7 +21,7 @@
   parent
 groupIdorg.apache.spark/groupId
 artifactIdspark-parent/artifactId
-version1.0.2/version
+version1.0.2-SNAPSHOT/version
 relativePath../../pom.xml/relativePath
   /parent
 

http://git-wip-us.apache.org/repos/asf/spark/blob/4be1dbd8/external/mqtt/pom.xml
--
diff --git a/external/mqtt/pom.xml b/external/mqtt/pom.xml
index ca022e0..ab5afa2 100644
--- a/external/mqtt/pom.xml

Git Push Summary

2014-07-25 Thread tdas
Repository: spark
Updated Tags:  refs/tags/v1.0.2-rc1 [deleted] 6053ff1be


git commit: [maven-release-plugin] prepare release v1.0.2-rc1

2014-07-25 Thread tdas
Repository: spark
Updated Branches:
  refs/heads/branch-1.0 4be1dbd82 - 8fb6f00e1


[maven-release-plugin] prepare release v1.0.2-rc1


Project: http://git-wip-us.apache.org/repos/asf/spark/repo
Commit: http://git-wip-us.apache.org/repos/asf/spark/commit/8fb6f00e
Tree: http://git-wip-us.apache.org/repos/asf/spark/tree/8fb6f00e
Diff: http://git-wip-us.apache.org/repos/asf/spark/diff/8fb6f00e

Branch: refs/heads/branch-1.0
Commit: 8fb6f00e195fb258f3f70f04756e07c259a2351f
Parents: 4be1dbd
Author: Tathagata Das tathagata.das1...@gmail.com
Authored: Fri Jul 25 21:21:15 2014 +
Committer: Tathagata Das tathagata.das1...@gmail.com
Committed: Fri Jul 25 21:21:15 2014 +

--
 assembly/pom.xml  | 2 +-
 bagel/pom.xml | 2 +-
 core/pom.xml  | 2 +-
 examples/pom.xml  | 2 +-
 external/flume/pom.xml| 2 +-
 external/kafka/pom.xml| 2 +-
 external/mqtt/pom.xml | 2 +-
 external/twitter/pom.xml  | 2 +-
 external/zeromq/pom.xml   | 2 +-
 extras/spark-ganglia-lgpl/pom.xml | 2 +-
 graphx/pom.xml| 2 +-
 mllib/pom.xml | 2 +-
 pom.xml   | 4 ++--
 repl/pom.xml  | 2 +-
 sql/catalyst/pom.xml  | 2 +-
 sql/core/pom.xml  | 2 +-
 sql/hive/pom.xml  | 2 +-
 streaming/pom.xml | 2 +-
 tools/pom.xml | 2 +-
 yarn/pom.xml  | 2 +-
 yarn/stable/pom.xml   | 2 +-
 21 files changed, 22 insertions(+), 22 deletions(-)
--


http://git-wip-us.apache.org/repos/asf/spark/blob/8fb6f00e/assembly/pom.xml
--
diff --git a/assembly/pom.xml b/assembly/pom.xml
index e290e79..6735379 100644
--- a/assembly/pom.xml
+++ b/assembly/pom.xml
@@ -21,7 +21,7 @@
   parent
 groupIdorg.apache.spark/groupId
 artifactIdspark-parent/artifactId
-version1.0.2-SNAPSHOT/version
+version1.0.2/version
 relativePath../pom.xml/relativePath
   /parent
 

http://git-wip-us.apache.org/repos/asf/spark/blob/8fb6f00e/bagel/pom.xml
--
diff --git a/bagel/pom.xml b/bagel/pom.xml
index c8ad40f..8a38b43 100644
--- a/bagel/pom.xml
+++ b/bagel/pom.xml
@@ -21,7 +21,7 @@
   parent
 groupIdorg.apache.spark/groupId
 artifactIdspark-parent/artifactId
-version1.0.2-SNAPSHOT/version
+version1.0.2/version
 relativePath../pom.xml/relativePath
   /parent
 

http://git-wip-us.apache.org/repos/asf/spark/blob/8fb6f00e/core/pom.xml
--
diff --git a/core/pom.xml b/core/pom.xml
index 2302d7b..23eea6e 100644
--- a/core/pom.xml
+++ b/core/pom.xml
@@ -21,7 +21,7 @@
   parent
 groupIdorg.apache.spark/groupId
 artifactIdspark-parent/artifactId
-version1.0.2-SNAPSHOT/version
+version1.0.2/version
 relativePath../pom.xml/relativePath
   /parent
 

http://git-wip-us.apache.org/repos/asf/spark/blob/8fb6f00e/examples/pom.xml
--
diff --git a/examples/pom.xml b/examples/pom.xml
index 9156a11..d158a75 100644
--- a/examples/pom.xml
+++ b/examples/pom.xml
@@ -21,7 +21,7 @@
   parent
 groupIdorg.apache.spark/groupId
 artifactIdspark-parent/artifactId
-version1.0.2-SNAPSHOT/version
+version1.0.2/version
 relativePath../pom.xml/relativePath
   /parent
 

http://git-wip-us.apache.org/repos/asf/spark/blob/8fb6f00e/external/flume/pom.xml
--
diff --git a/external/flume/pom.xml b/external/flume/pom.xml
index 1cefa15..f5f0d54 100644
--- a/external/flume/pom.xml
+++ b/external/flume/pom.xml
@@ -21,7 +21,7 @@
   parent
 groupIdorg.apache.spark/groupId
 artifactIdspark-parent/artifactId
-version1.0.2-SNAPSHOT/version
+version1.0.2/version
 relativePath../../pom.xml/relativePath
   /parent
 

http://git-wip-us.apache.org/repos/asf/spark/blob/8fb6f00e/external/kafka/pom.xml
--
diff --git a/external/kafka/pom.xml b/external/kafka/pom.xml
index cc05e69..8bc5c03 100644
--- a/external/kafka/pom.xml
+++ b/external/kafka/pom.xml
@@ -21,7 +21,7 @@
   parent
 groupIdorg.apache.spark/groupId
 artifactIdspark-parent/artifactId
-version1.0.2-SNAPSHOT/version
+version1.0.2/version
 relativePath../../pom.xml/relativePath
   /parent
 

http://git-wip-us.apache.org/repos/asf/spark/blob/8fb6f00e/external/mqtt/pom.xml
--
diff --git a/external/mqtt/pom.xml b/external/mqtt/pom.xml
index ab5afa2..ca022e0 100644
--- 

Git Push Summary

2014-07-25 Thread tdas
Repository: spark
Updated Tags:  refs/tags/v1.0.2-rc1 [created] 913e3de6d


git commit: [maven-release-plugin] prepare for next development iteration

2014-07-25 Thread tdas
Repository: spark
Updated Branches:
  refs/heads/branch-1.0 8fb6f00e1 - 03b684253


[maven-release-plugin] prepare for next development iteration


Project: http://git-wip-us.apache.org/repos/asf/spark/repo
Commit: http://git-wip-us.apache.org/repos/asf/spark/commit/03b68425
Tree: http://git-wip-us.apache.org/repos/asf/spark/tree/03b68425
Diff: http://git-wip-us.apache.org/repos/asf/spark/diff/03b68425

Branch: refs/heads/branch-1.0
Commit: 03b684253d550ee199323b787bea68bacb782689
Parents: 8fb6f00
Author: Tathagata Das tathagata.das1...@gmail.com
Authored: Fri Jul 25 21:21:21 2014 +
Committer: Tathagata Das tathagata.das1...@gmail.com
Committed: Fri Jul 25 21:21:21 2014 +

--
 assembly/pom.xml  | 2 +-
 bagel/pom.xml | 2 +-
 core/pom.xml  | 2 +-
 examples/pom.xml  | 2 +-
 external/flume/pom.xml| 2 +-
 external/kafka/pom.xml| 2 +-
 external/mqtt/pom.xml | 2 +-
 external/twitter/pom.xml  | 2 +-
 external/zeromq/pom.xml   | 2 +-
 extras/spark-ganglia-lgpl/pom.xml | 2 +-
 graphx/pom.xml| 2 +-
 mllib/pom.xml | 2 +-
 pom.xml   | 4 ++--
 repl/pom.xml  | 2 +-
 sql/catalyst/pom.xml  | 2 +-
 sql/core/pom.xml  | 2 +-
 sql/hive/pom.xml  | 2 +-
 streaming/pom.xml | 2 +-
 tools/pom.xml | 2 +-
 yarn/pom.xml  | 2 +-
 yarn/stable/pom.xml   | 2 +-
 21 files changed, 22 insertions(+), 22 deletions(-)
--


http://git-wip-us.apache.org/repos/asf/spark/blob/03b68425/assembly/pom.xml
--
diff --git a/assembly/pom.xml b/assembly/pom.xml
index 6735379..fbb993c 100644
--- a/assembly/pom.xml
+++ b/assembly/pom.xml
@@ -21,7 +21,7 @@
   parent
 groupIdorg.apache.spark/groupId
 artifactIdspark-parent/artifactId
-version1.0.2/version
+version1.0.3-SNAPSHOT/version
 relativePath../pom.xml/relativePath
   /parent
 

http://git-wip-us.apache.org/repos/asf/spark/blob/03b68425/bagel/pom.xml
--
diff --git a/bagel/pom.xml b/bagel/pom.xml
index 8a38b43..0f7b0d7 100644
--- a/bagel/pom.xml
+++ b/bagel/pom.xml
@@ -21,7 +21,7 @@
   parent
 groupIdorg.apache.spark/groupId
 artifactIdspark-parent/artifactId
-version1.0.2/version
+version1.0.3-SNAPSHOT/version
 relativePath../pom.xml/relativePath
   /parent
 

http://git-wip-us.apache.org/repos/asf/spark/blob/03b68425/core/pom.xml
--
diff --git a/core/pom.xml b/core/pom.xml
index 23eea6e..59110c2 100644
--- a/core/pom.xml
+++ b/core/pom.xml
@@ -21,7 +21,7 @@
   parent
 groupIdorg.apache.spark/groupId
 artifactIdspark-parent/artifactId
-version1.0.2/version
+version1.0.3-SNAPSHOT/version
 relativePath../pom.xml/relativePath
   /parent
 

http://git-wip-us.apache.org/repos/asf/spark/blob/03b68425/examples/pom.xml
--
diff --git a/examples/pom.xml b/examples/pom.xml
index d158a75..08a9ddd 100644
--- a/examples/pom.xml
+++ b/examples/pom.xml
@@ -21,7 +21,7 @@
   parent
 groupIdorg.apache.spark/groupId
 artifactIdspark-parent/artifactId
-version1.0.2/version
+version1.0.3-SNAPSHOT/version
 relativePath../pom.xml/relativePath
   /parent
 

http://git-wip-us.apache.org/repos/asf/spark/blob/03b68425/external/flume/pom.xml
--
diff --git a/external/flume/pom.xml b/external/flume/pom.xml
index f5f0d54..a486354 100644
--- a/external/flume/pom.xml
+++ b/external/flume/pom.xml
@@ -21,7 +21,7 @@
   parent
 groupIdorg.apache.spark/groupId
 artifactIdspark-parent/artifactId
-version1.0.2/version
+version1.0.3-SNAPSHOT/version
 relativePath../../pom.xml/relativePath
   /parent
 

http://git-wip-us.apache.org/repos/asf/spark/blob/03b68425/external/kafka/pom.xml
--
diff --git a/external/kafka/pom.xml b/external/kafka/pom.xml
index 8bc5c03..f68ad12 100644
--- a/external/kafka/pom.xml
+++ b/external/kafka/pom.xml
@@ -21,7 +21,7 @@
   parent
 groupIdorg.apache.spark/groupId
 artifactIdspark-parent/artifactId
-version1.0.2/version
+version1.0.3-SNAPSHOT/version
 relativePath../../pom.xml/relativePath
   /parent
 

http://git-wip-us.apache.org/repos/asf/spark/blob/03b68425/external/mqtt/pom.xml
--
diff --git a/external/mqtt/pom.xml b/external/mqtt/pom.xml
index ca022e0..b047b0f 100644
--- 

git commit: [SPARK-2103][Streaming] Change to ClassTag for KafkaInputDStream and fix reflection issue

2014-08-01 Thread tdas
Repository: spark
Updated Branches:
  refs/heads/master 284771efb - a32f0fb73


[SPARK-2103][Streaming] Change to ClassTag for KafkaInputDStream and fix 
reflection issue

This PR updates previous Manifest for KafkaInputDStream's Decoder to ClassTag, 
also fix the problem addressed in 
[SPARK-2103](https://issues.apache.org/jira/browse/SPARK-2103).

Previous Java interface cannot actually get the type of Decoder, so when using 
this Manifest to reconstruct the decode object will meet reflection exception.

Also for other two Java interfaces, ClassTag[String] is useless because calling 
Scala API will get the right implicit ClassTag.

Current Kafka unit test cannot actually verify the interface. I've tested these 
interfaces in my local and distribute settings.

Author: jerryshao saisai.s...@intel.com

Closes #1508 from jerryshao/SPARK-2103 and squashes the following commits:

e90c37b [jerryshao] Add Mima excludes
7529810 [jerryshao] Change Manifest to ClassTag for KafkaInputDStream's Decoder 
and fix Decoder construct issue when using Java API


Project: http://git-wip-us.apache.org/repos/asf/spark/repo
Commit: http://git-wip-us.apache.org/repos/asf/spark/commit/a32f0fb7
Tree: http://git-wip-us.apache.org/repos/asf/spark/tree/a32f0fb7
Diff: http://git-wip-us.apache.org/repos/asf/spark/diff/a32f0fb7

Branch: refs/heads/master
Commit: a32f0fb73a739c56208cafcd9f08618fb6dd8859
Parents: 284771e
Author: jerryshao saisai.s...@intel.com
Authored: Fri Aug 1 04:32:46 2014 -0700
Committer: Tathagata Das tathagata.das1...@gmail.com
Committed: Fri Aug 1 04:32:46 2014 -0700

--
 .../spark/streaming/kafka/KafkaInputDStream.scala   | 14 +++---
 .../apache/spark/streaming/kafka/KafkaUtils.scala   | 16 +---
 project/MimaExcludes.scala  |  7 ++-
 3 files changed, 18 insertions(+), 19 deletions(-)
--


http://git-wip-us.apache.org/repos/asf/spark/blob/a32f0fb7/external/kafka/src/main/scala/org/apache/spark/streaming/kafka/KafkaInputDStream.scala
--
diff --git 
a/external/kafka/src/main/scala/org/apache/spark/streaming/kafka/KafkaInputDStream.scala
 
b/external/kafka/src/main/scala/org/apache/spark/streaming/kafka/KafkaInputDStream.scala
index 38095e8..e20e2c8 100644
--- 
a/external/kafka/src/main/scala/org/apache/spark/streaming/kafka/KafkaInputDStream.scala
+++ 
b/external/kafka/src/main/scala/org/apache/spark/streaming/kafka/KafkaInputDStream.scala
@@ -18,7 +18,7 @@
 package org.apache.spark.streaming.kafka
 
 import scala.collection.Map
-import scala.reflect.ClassTag
+import scala.reflect.{classTag, ClassTag}
 
 import java.util.Properties
 import java.util.concurrent.Executors
@@ -48,8 +48,8 @@ private[streaming]
 class KafkaInputDStream[
   K: ClassTag,
   V: ClassTag,
-  U : Decoder[_]: Manifest,
-  T : Decoder[_]: Manifest](
+  U : Decoder[_]: ClassTag,
+  T : Decoder[_]: ClassTag](
 @transient ssc_ : StreamingContext,
 kafkaParams: Map[String, String],
 topics: Map[String, Int],
@@ -66,8 +66,8 @@ private[streaming]
 class KafkaReceiver[
   K: ClassTag,
   V: ClassTag,
-  U : Decoder[_]: Manifest,
-  T : Decoder[_]: Manifest](
+  U : Decoder[_]: ClassTag,
+  T : Decoder[_]: ClassTag](
 kafkaParams: Map[String, String],
 topics: Map[String, Int],
 storageLevel: StorageLevel
@@ -103,10 +103,10 @@ class KafkaReceiver[
   tryZookeeperConsumerGroupCleanup(zkConnect, kafkaParams(group.id))
 }
 
-val keyDecoder = 
manifest[U].runtimeClass.getConstructor(classOf[VerifiableProperties])
+val keyDecoder = 
classTag[U].runtimeClass.getConstructor(classOf[VerifiableProperties])
   .newInstance(consumerConfig.props)
   .asInstanceOf[Decoder[K]]
-val valueDecoder = 
manifest[T].runtimeClass.getConstructor(classOf[VerifiableProperties])
+val valueDecoder = 
classTag[T].runtimeClass.getConstructor(classOf[VerifiableProperties])
   .newInstance(consumerConfig.props)
   .asInstanceOf[Decoder[V]]
 

http://git-wip-us.apache.org/repos/asf/spark/blob/a32f0fb7/external/kafka/src/main/scala/org/apache/spark/streaming/kafka/KafkaUtils.scala
--
diff --git 
a/external/kafka/src/main/scala/org/apache/spark/streaming/kafka/KafkaUtils.scala
 
b/external/kafka/src/main/scala/org/apache/spark/streaming/kafka/KafkaUtils.scala
index 86bb91f..48668f7 100644
--- 
a/external/kafka/src/main/scala/org/apache/spark/streaming/kafka/KafkaUtils.scala
+++ 
b/external/kafka/src/main/scala/org/apache/spark/streaming/kafka/KafkaUtils.scala
@@ -65,7 +65,7 @@ object KafkaUtils {
*in its own thread.
* @param storageLevel Storage level to use for storing the received objects
*/
-  def createStream[K: ClassTag, V: ClassTag, U : Decoder[_]: Manifest, T : 

git commit: [SPARK-2379] Fix the bug that streaming's receiver may fall into a dead loop

2014-08-01 Thread tdas
Repository: spark
Updated Branches:
  refs/heads/master f5d9bea20 - b270309d7


[SPARK-2379] Fix the bug that streaming's receiver may fall into a dead loop

Author: joyyoj suns...@gmail.com

Closes #1694 from joyyoj/SPARK-2379 and squashes the following commits:

d73790d [joyyoj] SPARK-2379 Fix the bug that streaming's receiver may fall into 
a dead loop
22e7821 [joyyoj] Merge remote-tracking branch 'apache/master'
3f4a602 [joyyoj] Merge remote-tracking branch 'remotes/apache/master'
f4660c5 [joyyoj] [SPARK-1998] SparkFlumeEvent with body bigger than 1020 bytes 
are not read properly


Project: http://git-wip-us.apache.org/repos/asf/spark/repo
Commit: http://git-wip-us.apache.org/repos/asf/spark/commit/b270309d
Tree: http://git-wip-us.apache.org/repos/asf/spark/tree/b270309d
Diff: http://git-wip-us.apache.org/repos/asf/spark/diff/b270309d

Branch: refs/heads/master
Commit: b270309d7608fb749e402cd5afd36087446be398
Parents: f5d9bea
Author: joyyoj suns...@gmail.com
Authored: Fri Aug 1 13:41:55 2014 -0700
Committer: Tathagata Das tathagata.das1...@gmail.com
Committed: Fri Aug 1 13:41:55 2014 -0700

--
 .../org/apache/spark/streaming/receiver/ReceiverSupervisor.scala   | 2 +-
 1 file changed, 1 insertion(+), 1 deletion(-)
--


http://git-wip-us.apache.org/repos/asf/spark/blob/b270309d/streaming/src/main/scala/org/apache/spark/streaming/receiver/ReceiverSupervisor.scala
--
diff --git 
a/streaming/src/main/scala/org/apache/spark/streaming/receiver/ReceiverSupervisor.scala
 
b/streaming/src/main/scala/org/apache/spark/streaming/receiver/ReceiverSupervisor.scala
index 09be3a5..1f0244c 100644
--- 
a/streaming/src/main/scala/org/apache/spark/streaming/receiver/ReceiverSupervisor.scala
+++ 
b/streaming/src/main/scala/org/apache/spark/streaming/receiver/ReceiverSupervisor.scala
@@ -138,7 +138,7 @@ private[streaming] abstract class ReceiverSupervisor(
   onReceiverStop(message, error)
 } catch {
   case t: Throwable =
-stop(Error stopping receiver  + streamId, Some(t))
+logError(Error stopping receiver  + streamId + t.getStackTraceString)
 }
   }
 



git commit: [SPARK-2379] Fix the bug that streaming's receiver may fall into a dead loop

2014-08-01 Thread tdas
Repository: spark
Updated Branches:
  refs/heads/branch-1.0 886508d3b - 952e0d698


[SPARK-2379] Fix the bug that streaming's receiver may fall into a dead loop

Author: joyyoj suns...@gmail.com

Closes #1694 from joyyoj/SPARK-2379 and squashes the following commits:

d73790d [joyyoj] SPARK-2379 Fix the bug that streaming's receiver may fall into 
a dead loop
22e7821 [joyyoj] Merge remote-tracking branch 'apache/master'
3f4a602 [joyyoj] Merge remote-tracking branch 'remotes/apache/master'
f4660c5 [joyyoj] [SPARK-1998] SparkFlumeEvent with body bigger than 1020 bytes 
are not read properly

(cherry picked from commit b270309d7608fb749e402cd5afd36087446be398)
Signed-off-by: Tathagata Das tathagata.das1...@gmail.com


Project: http://git-wip-us.apache.org/repos/asf/spark/repo
Commit: http://git-wip-us.apache.org/repos/asf/spark/commit/952e0d69
Tree: http://git-wip-us.apache.org/repos/asf/spark/tree/952e0d69
Diff: http://git-wip-us.apache.org/repos/asf/spark/diff/952e0d69

Branch: refs/heads/branch-1.0
Commit: 952e0d69841b6218e7d1b8b23e7d74a4fcb1b381
Parents: 886508d
Author: joyyoj suns...@gmail.com
Authored: Fri Aug 1 13:41:55 2014 -0700
Committer: Tathagata Das tathagata.das1...@gmail.com
Committed: Fri Aug 1 13:42:16 2014 -0700

--
 .../org/apache/spark/streaming/receiver/ReceiverSupervisor.scala   | 2 +-
 1 file changed, 1 insertion(+), 1 deletion(-)
--


http://git-wip-us.apache.org/repos/asf/spark/blob/952e0d69/streaming/src/main/scala/org/apache/spark/streaming/receiver/ReceiverSupervisor.scala
--
diff --git 
a/streaming/src/main/scala/org/apache/spark/streaming/receiver/ReceiverSupervisor.scala
 
b/streaming/src/main/scala/org/apache/spark/streaming/receiver/ReceiverSupervisor.scala
index 09be3a5..1f0244c 100644
--- 
a/streaming/src/main/scala/org/apache/spark/streaming/receiver/ReceiverSupervisor.scala
+++ 
b/streaming/src/main/scala/org/apache/spark/streaming/receiver/ReceiverSupervisor.scala
@@ -138,7 +138,7 @@ private[streaming] abstract class ReceiverSupervisor(
   onReceiverStop(message, error)
 } catch {
   case t: Throwable =
-stop(Error stopping receiver  + streamId, Some(t))
+logError(Error stopping receiver  + streamId + t.getStackTraceString)
 }
   }
 



[1/2] [SPARK-1981] Add AWS Kinesis streaming support

2014-08-02 Thread tdas
Repository: spark
Updated Branches:
  refs/heads/master 67bd8e3c2 - 91f9504e6


http://git-wip-us.apache.org/repos/asf/spark/blob/91f9504e/extras/kinesis-asl/src/test/scala/org/apache/spark/streaming/kinesis/KinesisReceiverSuite.scala
--
diff --git 
a/extras/kinesis-asl/src/test/scala/org/apache/spark/streaming/kinesis/KinesisReceiverSuite.scala
 
b/extras/kinesis-asl/src/test/scala/org/apache/spark/streaming/kinesis/KinesisReceiverSuite.scala
new file mode 100644
index 000..41dbd64
--- /dev/null
+++ 
b/extras/kinesis-asl/src/test/scala/org/apache/spark/streaming/kinesis/KinesisReceiverSuite.scala
@@ -0,0 +1,275 @@
+/*
+ * Licensed to the Apache Software Foundation (ASF) under one or more
+ * contributor license agreements.  See the NOTICE file distributed with
+ * this work for additional information regarding copyright ownership.
+ * The ASF licenses this file to You under the Apache License, Version 2.0
+ * (the License); you may not use this file except in compliance with
+ * the License.  You may obtain a copy of the License at
+ *
+ *http://www.apache.org/licenses/LICENSE-2.0
+ *
+ * Unless required by applicable law or agreed to in writing, software
+ * distributed under the License is distributed on an AS IS BASIS,
+ * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+ * See the License for the specific language governing permissions and
+ * limitations under the License.
+ */
+package org.apache.spark.streaming.kinesis
+
+import java.nio.ByteBuffer
+
+import scala.collection.JavaConversions.seqAsJavaList
+
+import org.apache.spark.annotation.Experimental
+import org.apache.spark.storage.StorageLevel
+import org.apache.spark.streaming.Milliseconds
+import org.apache.spark.streaming.Seconds
+import org.apache.spark.streaming.StreamingContext
+import org.apache.spark.streaming.TestSuiteBase
+import org.apache.spark.streaming.util.Clock
+import org.apache.spark.streaming.util.ManualClock
+import org.scalatest.BeforeAndAfter
+import org.scalatest.Matchers
+import org.scalatest.mock.EasyMockSugar
+
+import 
com.amazonaws.services.kinesis.clientlibrary.exceptions.InvalidStateException
+import 
com.amazonaws.services.kinesis.clientlibrary.exceptions.KinesisClientLibDependencyException
+import 
com.amazonaws.services.kinesis.clientlibrary.exceptions.ShutdownException
+import 
com.amazonaws.services.kinesis.clientlibrary.exceptions.ThrottlingException
+import 
com.amazonaws.services.kinesis.clientlibrary.interfaces.IRecordProcessorCheckpointer
+import 
com.amazonaws.services.kinesis.clientlibrary.lib.worker.InitialPositionInStream
+import com.amazonaws.services.kinesis.clientlibrary.types.ShutdownReason
+import com.amazonaws.services.kinesis.model.Record
+
+/**
+ *  Suite of Kinesis streaming receiver tests focusing mostly on the 
KinesisRecordProcessor 
+ */
+class KinesisReceiverSuite extends TestSuiteBase with Matchers with 
BeforeAndAfter
+with EasyMockSugar {
+
+  val app = TestKinesisReceiver
+  val stream = mySparkStream
+  val endpoint = endpoint-url
+  val workerId = dummyWorkerId
+  val shardId = dummyShardId
+
+  val record1 = new Record()
+  record1.setData(ByteBuffer.wrap(Spark In Action.getBytes()))
+  val record2 = new Record()
+  record2.setData(ByteBuffer.wrap(Learning Spark.getBytes()))
+  val batch = List[Record](record1, record2)
+
+  var receiverMock: KinesisReceiver = _
+  var checkpointerMock: IRecordProcessorCheckpointer = _
+  var checkpointClockMock: ManualClock = _
+  var checkpointStateMock: KinesisCheckpointState = _
+  var currentClockMock: Clock = _
+
+  override def beforeFunction() = {
+receiverMock = mock[KinesisReceiver]
+checkpointerMock = mock[IRecordProcessorCheckpointer]
+checkpointClockMock = mock[ManualClock]
+checkpointStateMock = mock[KinesisCheckpointState]
+currentClockMock = mock[Clock]
+  }
+
+  test(kinesis utils api) {
+val ssc = new StreamingContext(master, framework, batchDuration)
+// Tests the API, does not actually test data receiving
+val kinesisStream = KinesisUtils.createStream(ssc, mySparkStream,
+  https://kinesis.us-west-2.amazonaws.com;, Seconds(2),
+  InitialPositionInStream.LATEST, StorageLevel.MEMORY_AND_DISK_2);
+ssc.stop()
+  }
+
+  test(process records including store and checkpoint) {
+val expectedCheckpointIntervalMillis = 10
+expecting {
+  receiverMock.isStopped().andReturn(false).once()
+  receiverMock.store(record1.getData().array()).once()
+  receiverMock.store(record2.getData().array()).once()
+  checkpointStateMock.shouldCheckpoint().andReturn(true).once()
+  checkpointerMock.checkpoint().once()
+  checkpointStateMock.advanceCheckpoint().once()
+}
+whenExecuting(receiverMock, checkpointerMock, checkpointStateMock) {
+  val recordProcessor = new KinesisRecordProcessor(receiverMock, workerId,
+  checkpointStateMock)
+  

[2/2] git commit: [SPARK-1981] Add AWS Kinesis streaming support

2014-08-02 Thread tdas
[SPARK-1981] Add AWS Kinesis streaming support

Author: Chris Fregly ch...@fregly.com

Closes #1434 from cfregly/master and squashes the following commits:

4774581 [Chris Fregly] updated docs, renamed retry to retryRandom to be more 
clear, removed retries around store() method
0393795 [Chris Fregly] moved Kinesis examples out of examples/ and back into 
extras/kinesis-asl
691a6be [Chris Fregly] fixed tests and formatting, fixed a bug with 
JavaKinesisWordCount during union of streams
0e1c67b [Chris Fregly] Merge remote-tracking branch 'upstream/master'
74e5c7c [Chris Fregly] updated per TD's feedback.  simplified examples, updated 
docs
e33cbeb [Chris Fregly] Merge remote-tracking branch 'upstream/master'
bf614e9 [Chris Fregly] per matei's feedback:  moved the kinesis examples into 
the examples/ dir
d17ca6d [Chris Fregly] per TD's feedback:  updated docs, simplified the 
KinesisUtils api
912640c [Chris Fregly] changed the foundKinesis class to be a publically-avail 
class
db3eefd [Chris Fregly] Merge remote-tracking branch 'upstream/master'
21de67f [Chris Fregly] Merge remote-tracking branch 'upstream/master'
6c39561 [Chris Fregly] parameterized the versions of the aws java sdk and 
kinesis client
338997e [Chris Fregly] improve build docs for kinesis
828f8ae [Chris Fregly] more cleanup
e7c8978 [Chris Fregly] Merge remote-tracking branch 'upstream/master'
cd68c0d [Chris Fregly] fixed typos and backward compatibility
d18e680 [Chris Fregly] Merge remote-tracking branch 'upstream/master'
b3b0ff1 [Chris Fregly] [SPARK-1981] Add AWS Kinesis streaming support


Project: http://git-wip-us.apache.org/repos/asf/spark/repo
Commit: http://git-wip-us.apache.org/repos/asf/spark/commit/91f9504e
Tree: http://git-wip-us.apache.org/repos/asf/spark/tree/91f9504e
Diff: http://git-wip-us.apache.org/repos/asf/spark/diff/91f9504e

Branch: refs/heads/master
Commit: 91f9504e6086fac05b40545099f9818949c24bca
Parents: 67bd8e3
Author: Chris Fregly ch...@fregly.com
Authored: Sat Aug 2 13:35:35 2014 -0700
Committer: Tathagata Das tathagata.das1...@gmail.com
Committed: Sat Aug 2 13:35:35 2014 -0700

--
 bin/run-example |   3 +-
 bin/run-example2.cmd|   3 +-
 dev/audit-release/audit_release.py  |   4 +-
 .../sbt_app_core/src/main/scala/SparkApp.scala  |   7 +
 dev/audit-release/sbt_app_kinesis/build.sbt |  28 ++
 .../src/main/scala/SparkApp.scala   |  33 +++
 dev/create-release/create-release.sh|   4 +-
 dev/run-tests   |   3 +
 docs/streaming-custom-receivers.md  |   4 +-
 docs/streaming-kinesis.md   |  58 
 docs/streaming-programming-guide.md |  12 +-
 examples/pom.xml|  13 +
 extras/kinesis-asl/pom.xml  |  96 +++
 .../streaming/JavaKinesisWordCountASL.java  | 180 
 .../src/main/resources/log4j.properties |  37 +++
 .../streaming/KinesisWordCountASL.scala | 251 +
 .../kinesis/KinesisCheckpointState.scala|  56 
 .../streaming/kinesis/KinesisReceiver.scala | 149 ++
 .../kinesis/KinesisRecordProcessor.scala| 212 ++
 .../spark/streaming/kinesis/KinesisUtils.scala  |  96 +++
 .../kinesis/JavaKinesisStreamSuite.java |  41 +++
 .../src/test/resources/log4j.properties |  26 ++
 .../kinesis/KinesisReceiverSuite.scala  | 275 +++
 pom.xml |  10 +
 project/SparkBuild.scala|   6 +-
 25 files changed, 1592 insertions(+), 15 deletions(-)
--


http://git-wip-us.apache.org/repos/asf/spark/blob/91f9504e/bin/run-example
--
diff --git a/bin/run-example b/bin/run-example
index 942706d..68a3570 100755
--- a/bin/run-example
+++ b/bin/run-example
@@ -29,7 +29,8 @@ if [ -n $1 ]; then
 else
   echo Usage: ./bin/run-example example-class [example-args] 12
   echo   - set MASTER=XX to use a specific master 12
-  echo   - can use abbreviated example class name (e.g. SparkPi, 
mllib.LinearRegression) 12
+  echo   - can use abbreviated example class name relative to 
com.apache.spark.examples 12
+  echo  (e.g. SparkPi, mllib.LinearRegression, 
streaming.KinesisWordCountASL) 12
   exit 1
 fi
 

http://git-wip-us.apache.org/repos/asf/spark/blob/91f9504e/bin/run-example2.cmd
--
diff --git a/bin/run-example2.cmd b/bin/run-example2.cmd
index eadedd7..b29bf90 100644
--- a/bin/run-example2.cmd
+++ b/bin/run-example2.cmd
@@ -32,7 +32,8 @@ rem Test that an argument was given
 if not x%1==x goto arg_given
   echo Usage: run-example ^example-class^ [example-args]
   echo   - 

[2/2] git commit: [SPARK-1981] Add AWS Kinesis streaming support

2014-08-02 Thread tdas
[SPARK-1981] Add AWS Kinesis streaming support

Author: Chris Fregly ch...@fregly.com

Closes #1434 from cfregly/master and squashes the following commits:

4774581 [Chris Fregly] updated docs, renamed retry to retryRandom to be more 
clear, removed retries around store() method
0393795 [Chris Fregly] moved Kinesis examples out of examples/ and back into 
extras/kinesis-asl
691a6be [Chris Fregly] fixed tests and formatting, fixed a bug with 
JavaKinesisWordCount during union of streams
0e1c67b [Chris Fregly] Merge remote-tracking branch 'upstream/master'
74e5c7c [Chris Fregly] updated per TD's feedback.  simplified examples, updated 
docs
e33cbeb [Chris Fregly] Merge remote-tracking branch 'upstream/master'
bf614e9 [Chris Fregly] per matei's feedback:  moved the kinesis examples into 
the examples/ dir
d17ca6d [Chris Fregly] per TD's feedback:  updated docs, simplified the 
KinesisUtils api
912640c [Chris Fregly] changed the foundKinesis class to be a publically-avail 
class
db3eefd [Chris Fregly] Merge remote-tracking branch 'upstream/master'
21de67f [Chris Fregly] Merge remote-tracking branch 'upstream/master'
6c39561 [Chris Fregly] parameterized the versions of the aws java sdk and 
kinesis client
338997e [Chris Fregly] improve build docs for kinesis
828f8ae [Chris Fregly] more cleanup
e7c8978 [Chris Fregly] Merge remote-tracking branch 'upstream/master'
cd68c0d [Chris Fregly] fixed typos and backward compatibility
d18e680 [Chris Fregly] Merge remote-tracking branch 'upstream/master'
b3b0ff1 [Chris Fregly] [SPARK-1981] Add AWS Kinesis streaming support

(cherry picked from commit 91f9504e6086fac05b40545099f9818949c24bca)
Signed-off-by: Tathagata Das tathagata.das1...@gmail.com


Project: http://git-wip-us.apache.org/repos/asf/spark/repo
Commit: http://git-wip-us.apache.org/repos/asf/spark/commit/bb0ac6d7
Tree: http://git-wip-us.apache.org/repos/asf/spark/tree/bb0ac6d7
Diff: http://git-wip-us.apache.org/repos/asf/spark/diff/bb0ac6d7

Branch: refs/heads/branch-1.1
Commit: bb0ac6d7c91c491a99c252e6cb4aea40efe9b190
Parents: 91de0dc
Author: Chris Fregly ch...@fregly.com
Authored: Sat Aug 2 13:35:35 2014 -0700
Committer: Tathagata Das tathagata.das1...@gmail.com
Committed: Sat Aug 2 13:35:57 2014 -0700

--
 bin/run-example |   3 +-
 bin/run-example2.cmd|   3 +-
 dev/audit-release/audit_release.py  |   4 +-
 .../sbt_app_core/src/main/scala/SparkApp.scala  |   7 +
 dev/audit-release/sbt_app_kinesis/build.sbt |  28 ++
 .../src/main/scala/SparkApp.scala   |  33 +++
 dev/create-release/create-release.sh|   4 +-
 dev/run-tests   |   3 +
 docs/streaming-custom-receivers.md  |   4 +-
 docs/streaming-kinesis.md   |  58 
 docs/streaming-programming-guide.md |  12 +-
 examples/pom.xml|  13 +
 extras/kinesis-asl/pom.xml  |  96 +++
 .../streaming/JavaKinesisWordCountASL.java  | 180 
 .../src/main/resources/log4j.properties |  37 +++
 .../streaming/KinesisWordCountASL.scala | 251 +
 .../kinesis/KinesisCheckpointState.scala|  56 
 .../streaming/kinesis/KinesisReceiver.scala | 149 ++
 .../kinesis/KinesisRecordProcessor.scala| 212 ++
 .../spark/streaming/kinesis/KinesisUtils.scala  |  96 +++
 .../kinesis/JavaKinesisStreamSuite.java |  41 +++
 .../src/test/resources/log4j.properties |  26 ++
 .../kinesis/KinesisReceiverSuite.scala  | 275 +++
 pom.xml |  10 +
 project/SparkBuild.scala|   6 +-
 25 files changed, 1592 insertions(+), 15 deletions(-)
--


http://git-wip-us.apache.org/repos/asf/spark/blob/bb0ac6d7/bin/run-example
--
diff --git a/bin/run-example b/bin/run-example
index 942706d..68a3570 100755
--- a/bin/run-example
+++ b/bin/run-example
@@ -29,7 +29,8 @@ if [ -n $1 ]; then
 else
   echo Usage: ./bin/run-example example-class [example-args] 12
   echo   - set MASTER=XX to use a specific master 12
-  echo   - can use abbreviated example class name (e.g. SparkPi, 
mllib.LinearRegression) 12
+  echo   - can use abbreviated example class name relative to 
com.apache.spark.examples 12
+  echo  (e.g. SparkPi, mllib.LinearRegression, 
streaming.KinesisWordCountASL) 12
   exit 1
 fi
 

http://git-wip-us.apache.org/repos/asf/spark/blob/bb0ac6d7/bin/run-example2.cmd
--
diff --git a/bin/run-example2.cmd b/bin/run-example2.cmd
index eadedd7..b29bf90 100644
--- a/bin/run-example2.cmd
+++ b/bin/run-example2.cmd
@@ -32,7 +32,8 @@ rem 

[1/2] [SPARK-1981] Add AWS Kinesis streaming support

2014-08-02 Thread tdas
Repository: spark
Updated Branches:
  refs/heads/branch-1.1 91de0dc16 - bb0ac6d7c


http://git-wip-us.apache.org/repos/asf/spark/blob/bb0ac6d7/extras/kinesis-asl/src/test/scala/org/apache/spark/streaming/kinesis/KinesisReceiverSuite.scala
--
diff --git 
a/extras/kinesis-asl/src/test/scala/org/apache/spark/streaming/kinesis/KinesisReceiverSuite.scala
 
b/extras/kinesis-asl/src/test/scala/org/apache/spark/streaming/kinesis/KinesisReceiverSuite.scala
new file mode 100644
index 000..41dbd64
--- /dev/null
+++ 
b/extras/kinesis-asl/src/test/scala/org/apache/spark/streaming/kinesis/KinesisReceiverSuite.scala
@@ -0,0 +1,275 @@
+/*
+ * Licensed to the Apache Software Foundation (ASF) under one or more
+ * contributor license agreements.  See the NOTICE file distributed with
+ * this work for additional information regarding copyright ownership.
+ * The ASF licenses this file to You under the Apache License, Version 2.0
+ * (the License); you may not use this file except in compliance with
+ * the License.  You may obtain a copy of the License at
+ *
+ *http://www.apache.org/licenses/LICENSE-2.0
+ *
+ * Unless required by applicable law or agreed to in writing, software
+ * distributed under the License is distributed on an AS IS BASIS,
+ * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+ * See the License for the specific language governing permissions and
+ * limitations under the License.
+ */
+package org.apache.spark.streaming.kinesis
+
+import java.nio.ByteBuffer
+
+import scala.collection.JavaConversions.seqAsJavaList
+
+import org.apache.spark.annotation.Experimental
+import org.apache.spark.storage.StorageLevel
+import org.apache.spark.streaming.Milliseconds
+import org.apache.spark.streaming.Seconds
+import org.apache.spark.streaming.StreamingContext
+import org.apache.spark.streaming.TestSuiteBase
+import org.apache.spark.streaming.util.Clock
+import org.apache.spark.streaming.util.ManualClock
+import org.scalatest.BeforeAndAfter
+import org.scalatest.Matchers
+import org.scalatest.mock.EasyMockSugar
+
+import 
com.amazonaws.services.kinesis.clientlibrary.exceptions.InvalidStateException
+import 
com.amazonaws.services.kinesis.clientlibrary.exceptions.KinesisClientLibDependencyException
+import 
com.amazonaws.services.kinesis.clientlibrary.exceptions.ShutdownException
+import 
com.amazonaws.services.kinesis.clientlibrary.exceptions.ThrottlingException
+import 
com.amazonaws.services.kinesis.clientlibrary.interfaces.IRecordProcessorCheckpointer
+import 
com.amazonaws.services.kinesis.clientlibrary.lib.worker.InitialPositionInStream
+import com.amazonaws.services.kinesis.clientlibrary.types.ShutdownReason
+import com.amazonaws.services.kinesis.model.Record
+
+/**
+ *  Suite of Kinesis streaming receiver tests focusing mostly on the 
KinesisRecordProcessor 
+ */
+class KinesisReceiverSuite extends TestSuiteBase with Matchers with 
BeforeAndAfter
+with EasyMockSugar {
+
+  val app = TestKinesisReceiver
+  val stream = mySparkStream
+  val endpoint = endpoint-url
+  val workerId = dummyWorkerId
+  val shardId = dummyShardId
+
+  val record1 = new Record()
+  record1.setData(ByteBuffer.wrap(Spark In Action.getBytes()))
+  val record2 = new Record()
+  record2.setData(ByteBuffer.wrap(Learning Spark.getBytes()))
+  val batch = List[Record](record1, record2)
+
+  var receiverMock: KinesisReceiver = _
+  var checkpointerMock: IRecordProcessorCheckpointer = _
+  var checkpointClockMock: ManualClock = _
+  var checkpointStateMock: KinesisCheckpointState = _
+  var currentClockMock: Clock = _
+
+  override def beforeFunction() = {
+receiverMock = mock[KinesisReceiver]
+checkpointerMock = mock[IRecordProcessorCheckpointer]
+checkpointClockMock = mock[ManualClock]
+checkpointStateMock = mock[KinesisCheckpointState]
+currentClockMock = mock[Clock]
+  }
+
+  test(kinesis utils api) {
+val ssc = new StreamingContext(master, framework, batchDuration)
+// Tests the API, does not actually test data receiving
+val kinesisStream = KinesisUtils.createStream(ssc, mySparkStream,
+  https://kinesis.us-west-2.amazonaws.com;, Seconds(2),
+  InitialPositionInStream.LATEST, StorageLevel.MEMORY_AND_DISK_2);
+ssc.stop()
+  }
+
+  test(process records including store and checkpoint) {
+val expectedCheckpointIntervalMillis = 10
+expecting {
+  receiverMock.isStopped().andReturn(false).once()
+  receiverMock.store(record1.getData().array()).once()
+  receiverMock.store(record2.getData().array()).once()
+  checkpointStateMock.shouldCheckpoint().andReturn(true).once()
+  checkpointerMock.checkpoint().once()
+  checkpointStateMock.advanceCheckpoint().once()
+}
+whenExecuting(receiverMock, checkpointerMock, checkpointStateMock) {
+  val recordProcessor = new KinesisRecordProcessor(receiverMock, workerId,
+  checkpointStateMock)
+  

svn commit: r6045 - /dev/spark/spark-1.0.2-rc1/

2014-08-04 Thread tdas
Author: tdas
Date: Tue Aug  5 00:19:19 2014
New Revision: 6045

Log:
Adding spark-1.0.2-rc1

Added:
dev/spark/spark-1.0.2-rc1/
dev/spark/spark-1.0.2-rc1/spark-1.0.2-bin-cdh4.tgz   (with props)
dev/spark/spark-1.0.2-rc1/spark-1.0.2-bin-cdh4.tgz.asc
dev/spark/spark-1.0.2-rc1/spark-1.0.2-bin-cdh4.tgz.md5
dev/spark/spark-1.0.2-rc1/spark-1.0.2-bin-cdh4.tgz.sha
dev/spark/spark-1.0.2-rc1/spark-1.0.2-bin-hadoop1.tgz   (with props)
dev/spark/spark-1.0.2-rc1/spark-1.0.2-bin-hadoop1.tgz.asc
dev/spark/spark-1.0.2-rc1/spark-1.0.2-bin-hadoop1.tgz.md5
dev/spark/spark-1.0.2-rc1/spark-1.0.2-bin-hadoop1.tgz.sha
dev/spark/spark-1.0.2-rc1/spark-1.0.2-bin-hadoop2.tgz   (with props)
dev/spark/spark-1.0.2-rc1/spark-1.0.2-bin-hadoop2.tgz.asc
dev/spark/spark-1.0.2-rc1/spark-1.0.2-bin-hadoop2.tgz.md5
dev/spark/spark-1.0.2-rc1/spark-1.0.2-bin-hadoop2.tgz.sha
dev/spark/spark-1.0.2-rc1/spark-1.0.2.tgz   (with props)
dev/spark/spark-1.0.2-rc1/spark-1.0.2.tgz.asc
dev/spark/spark-1.0.2-rc1/spark-1.0.2.tgz.md5
dev/spark/spark-1.0.2-rc1/spark-1.0.2.tgz.sha

Added: dev/spark/spark-1.0.2-rc1/spark-1.0.2-bin-cdh4.tgz
==
Binary file - no diff available.

Propchange: dev/spark/spark-1.0.2-rc1/spark-1.0.2-bin-cdh4.tgz
--
svn:mime-type = application/octet-stream

Added: dev/spark/spark-1.0.2-rc1/spark-1.0.2-bin-cdh4.tgz.asc
==
--- dev/spark/spark-1.0.2-rc1/spark-1.0.2-bin-cdh4.tgz.asc (added)
+++ dev/spark/spark-1.0.2-rc1/spark-1.0.2-bin-cdh4.tgz.asc Tue Aug  5 00:19:19 
2014
@@ -0,0 +1,17 @@
+-BEGIN PGP SIGNATURE-
+Version: GnuPG v1.4.14 (GNU/Linux)
+
+iQIcBAABAgAGBQJT0tdUAAoJEF2VHP+H/RqXKDYQAJd5kz96LmBnpeGCx24JtSf2
+0vR4gTRZIYX2Er+PVY0GAKNjzTI8rij1aK7xGQmCsFvAmW0EYbXd7r5kyCqip+0d
+Us7A8ultU9R1BPo2XaQgJSPg16xnT52ogSPXSOV40kLOyeWnQxiDL/O7/zThddjr
+FHwMVvaS55dJOBzhqCPqFBV+19nqKgqIJ/j6PPvFhAIdI7NfMnb1cK0dyLeQEuG8
+uFTky8hNDbhYMskegKmShhsnN/Ub1DvYAAbinBdOIH1zJnyNjnyRjDYRUP+K0Hu5
+VThuQdeNHb7lwvktJgaxAjbsguAYJiRq+QA7yfj2QteB7E4ruGHZ8x48QcDKWYpv
+4MDBfYRa4jjVX3c0rC0ZItfdErS54091z9sQIjIqTEk3pEr+AHOoptuBPEQKY3Ge
+fKAJdqCViJtG+YU7vc5hfvQ92tg3kXAiHk0rAlat458Mag+NecJaCAWcDugrVxIw
+V4qOWuih4nZRpRn1mZFnMXM6gkOxq2CZSwagX8hz7xRFrIO4SlgqYWuzX3M86OvS
+ZAbaTL6loTOigA49AuW0ffBdgSf72Od0l9QSgu7Z9ufsiPeaqtUC5jXn6PLWw2gI
+ZqtNSt2NNe7OOgXDcCOWjx1PcD+PKgI/99NJvhxeAv9H7zvsjqoV19eqsFqdP37W
+6RMNUXNNeq5KnKqkS+Jk
+=o9VB
+-END PGP SIGNATURE-

Added: dev/spark/spark-1.0.2-rc1/spark-1.0.2-bin-cdh4.tgz.md5
==
--- dev/spark/spark-1.0.2-rc1/spark-1.0.2-bin-cdh4.tgz.md5 (added)
+++ dev/spark/spark-1.0.2-rc1/spark-1.0.2-bin-cdh4.tgz.md5 Tue Aug  5 00:19:19 
2014
@@ -0,0 +1 @@
+spark-1.0.2-bin-cdh4.tgz: DF C2 27 27 81 24 7A 24  1F 1F 58 78 44 3A 5F 19

Added: dev/spark/spark-1.0.2-rc1/spark-1.0.2-bin-cdh4.tgz.sha
==
--- dev/spark/spark-1.0.2-rc1/spark-1.0.2-bin-cdh4.tgz.sha (added)
+++ dev/spark/spark-1.0.2-rc1/spark-1.0.2-bin-cdh4.tgz.sha Tue Aug  5 00:19:19 
2014
@@ -0,0 +1,3 @@
+spark-1.0.2-bin-cdh4.tgz: EC221342 C52DEBF6 5A52BBC9 6CCD1532 6FA4F177 37D09DD0
+  E9F67B45 2174B721 F47ADE98 8F3CD8F0 008591F7 318C2856
+  0B3B224D D2E254CD F67790DC 0A136778

Added: dev/spark/spark-1.0.2-rc1/spark-1.0.2-bin-hadoop1.tgz
==
Binary file - no diff available.

Propchange: dev/spark/spark-1.0.2-rc1/spark-1.0.2-bin-hadoop1.tgz
--
svn:mime-type = application/octet-stream

Added: dev/spark/spark-1.0.2-rc1/spark-1.0.2-bin-hadoop1.tgz.asc
==
--- dev/spark/spark-1.0.2-rc1/spark-1.0.2-bin-hadoop1.tgz.asc (added)
+++ dev/spark/spark-1.0.2-rc1/spark-1.0.2-bin-hadoop1.tgz.asc Tue Aug  5 
00:19:19 2014
@@ -0,0 +1,17 @@
+-BEGIN PGP SIGNATURE-
+Version: GnuPG v1.4.14 (GNU/Linux)
+
+iQIcBAABAgAGBQJT0tQsAAoJEF2VHP+H/RqXKKcP/34Q/wJqOonIyivWSHTsMl79
+EClGDar3jmJC0p3yPBdl6jZ+JVwkwTqlvHfeczM0POTGg0EeEcTJuk3wgB85FF6N
+oHlKraQ4Rf0LEt1SmcrmtQLOOkrajWJWBH9VAXnGke9k2Kv9N5uu123l/b0jPONO
+IVd09fdsBEK2lVYKi7Eyh3NQ4riDwPzk7ExDMHvSA+Ekp5V/zRSqgBSVOeziajAo
+a2NhoJHaLHMhghL3zmqjJUH68Dyde8iYvJStrWSoHZJ0C+V14Lj0ASD/SA6wfnYO
+KOKD1y2GTUlzoMpju5so90Dlap+0Rh2hobbZ2jv1mry+NQAwTUULFVCascBiq0mB
+9HAaRmMRR/I4HPXk0JoTYHXdl6cgFBDms8sSKLEB2GLw06AqmhIUSUJ+hRvAPCqV
+MJGsqzjn6wU9YvcinMjmdbPPBHYg3UdB/nT5F9TesbLSw0d7rEfpjMr2smpcM3Cw
+klwhlp3o7R2mEkdLD2P9auBYbfY3Z7FEzPKQXp118i5CuTmfa4Kaq0bbhk8Iv8sH
+GbaoKVFcJt3z95OTrX5N4Hy1Rn9xkwYaj5gal08OmUWsfuRxF9tb2+E5iu0uw4r0

Git Push Summary

2014-08-04 Thread tdas
Repository: spark
Updated Tags:  refs/tags/v0.7.0-bizo-1 [created] 02b4a0d96
  refs/tags/v0.9.1-rc1 [created] 20a2d247a
  refs/tags/v0.9.1-rc2 [created] 95d70498b
  refs/tags/v0.9.1-rc3 [created] 931ac8a0e
  refs/tags/v1.0.0-rc1 [created] c8d0eb980
  refs/tags/v1.0.0-rc10 [created] eafc6f6f4
  refs/tags/v1.0.0-rc2 [created] 327ab1e24
  refs/tags/v1.0.0-rc3 [created] e7c46933a
  refs/tags/v1.0.0-rc5 [created] 04ce12153
  refs/tags/v1.0.0-rc6 [created] aab03f5f9
  refs/tags/v1.0.0-rc7 [created] bc81ecfdc
  refs/tags/v1.0.0-rc8 [created] 9abe0192a
  refs/tags/v1.0.0-rc9 [created] 093aeb40f
  refs/tags/v1.0.1-rc1 [created] 2a9114542
  refs/tags/v1.0.1-rc2 [created] 9c4493512
  refs/tags/v1.0.2 [created] 8fb6f00e1

-
To unsubscribe, e-mail: commits-unsubscr...@spark.apache.org
For additional commands, e-mail: commits-h...@spark.apache.org



Git Push Summary

2014-08-04 Thread tdas
Repository: spark
Updated Tags:  refs/tags/v1.0.0-rc2 [deleted] 327ab1e24

-
To unsubscribe, e-mail: commits-unsubscr...@spark.apache.org
For additional commands, e-mail: commits-h...@spark.apache.org



Git Push Summary

2014-08-04 Thread tdas
Repository: spark
Updated Tags:  refs/tags/v1.0.0-rc7 [deleted] bc81ecfdc

-
To unsubscribe, e-mail: commits-unsubscr...@spark.apache.org
For additional commands, e-mail: commits-h...@spark.apache.org



Git Push Summary

2014-08-04 Thread tdas
Repository: spark
Updated Tags:  refs/tags/v1.0.0-rc6 [deleted] aab03f5f9

-
To unsubscribe, e-mail: commits-unsubscr...@spark.apache.org
For additional commands, e-mail: commits-h...@spark.apache.org



Git Push Summary

2014-08-04 Thread tdas
Repository: spark
Updated Tags:  refs/tags/v1.0.0-rc8 [deleted] 9abe0192a

-
To unsubscribe, e-mail: commits-unsubscr...@spark.apache.org
For additional commands, e-mail: commits-h...@spark.apache.org



Git Push Summary

2014-08-04 Thread tdas
Repository: spark
Updated Tags:  refs/tags/v1.0.1-rc1 [deleted] 2a9114542

-
To unsubscribe, e-mail: commits-unsubscr...@spark.apache.org
For additional commands, e-mail: commits-h...@spark.apache.org



Git Push Summary

2014-08-04 Thread tdas
Repository: spark
Updated Tags:  refs/tags/v1.0.0-rc10 [deleted] eafc6f6f4

-
To unsubscribe, e-mail: commits-unsubscr...@spark.apache.org
For additional commands, e-mail: commits-h...@spark.apache.org



Git Push Summary

2014-08-04 Thread tdas
Repository: spark
Updated Tags:  refs/tags/v0.9.2-rc1 [deleted] 5a3020615

-
To unsubscribe, e-mail: commits-unsubscr...@spark.apache.org
For additional commands, e-mail: commits-h...@spark.apache.org



Git Push Summary

2014-08-04 Thread tdas
Repository: spark
Updated Tags:  refs/tags/v0.9.1-rc3 [deleted] 931ac8a0e

-
To unsubscribe, e-mail: commits-unsubscr...@spark.apache.org
For additional commands, e-mail: commits-h...@spark.apache.org



Git Push Summary

2014-08-04 Thread tdas
Repository: spark
Updated Tags:  refs/tags/v0.9.1-rc1 [deleted] 20a2d247a

-
To unsubscribe, e-mail: commits-unsubscr...@spark.apache.org
For additional commands, e-mail: commits-h...@spark.apache.org



Git Push Summary

2014-08-04 Thread tdas
Repository: spark
Updated Tags:  refs/tags/v0.9.1-rc2 [deleted] 95d70498b

-
To unsubscribe, e-mail: commits-unsubscr...@spark.apache.org
For additional commands, e-mail: commits-h...@spark.apache.org



Git Push Summary

2014-08-04 Thread tdas
Repository: spark
Updated Tags:  refs/tags/v1.0.2-rc1 [deleted] 913e3de6d

-
To unsubscribe, e-mail: commits-unsubscr...@spark.apache.org
For additional commands, e-mail: commits-h...@spark.apache.org



git commit: Updating version for Spark 1.0.3 development.

2014-08-04 Thread tdas
Repository: spark
Updated Branches:
  refs/heads/branch-1.0 439cdc581 - 3fbe3447e


Updating version for Spark 1.0.3 development.


Project: http://git-wip-us.apache.org/repos/asf/spark/repo
Commit: http://git-wip-us.apache.org/repos/asf/spark/commit/3fbe3447
Tree: http://git-wip-us.apache.org/repos/asf/spark/tree/3fbe3447
Diff: http://git-wip-us.apache.org/repos/asf/spark/diff/3fbe3447

Branch: refs/heads/branch-1.0
Commit: 3fbe3447ec81b753985753dd47ad9941c78965e7
Parents: 439cdc5
Author: Tathagata Das tathagata.das1...@gmail.com
Authored: Mon Aug 4 18:45:18 2014 -0700
Committer: Tathagata Das tathagata.das1...@gmail.com
Committed: Mon Aug 4 18:45:18 2014 -0700

--
 core/src/main/scala/org/apache/spark/SparkContext.scala| 2 +-
 docs/_config.yml   | 4 ++--
 extras/java8-tests/pom.xml | 2 +-
 project/SparkBuild.scala   | 2 +-
 project/project/SparkPluginBuild.scala | 2 +-
 python/epydoc.conf | 2 +-
 python/pyspark/shell.py| 2 +-
 repl/src/main/scala/org/apache/spark/repl/SparkILoopInit.scala | 2 +-
 yarn/alpha/pom.xml | 2 +-
 9 files changed, 10 insertions(+), 10 deletions(-)
--


http://git-wip-us.apache.org/repos/asf/spark/blob/3fbe3447/core/src/main/scala/org/apache/spark/SparkContext.scala
--
diff --git a/core/src/main/scala/org/apache/spark/SparkContext.scala 
b/core/src/main/scala/org/apache/spark/SparkContext.scala
index 113c691..4a7d33d 100644
--- a/core/src/main/scala/org/apache/spark/SparkContext.scala
+++ b/core/src/main/scala/org/apache/spark/SparkContext.scala
@@ -1265,7 +1265,7 @@ class SparkContext(config: SparkConf) extends Logging {
  */
 object SparkContext extends Logging {
 
-  private[spark] val SPARK_VERSION = 1.0.2
+  private[spark] val SPARK_VERSION = 1.0.3-SNAPSHOT
 
   private[spark] val SPARK_JOB_DESCRIPTION = spark.job.description
 

http://git-wip-us.apache.org/repos/asf/spark/blob/3fbe3447/docs/_config.yml
--
diff --git a/docs/_config.yml b/docs/_config.yml
index 491b935..05a04fd 100644
--- a/docs/_config.yml
+++ b/docs/_config.yml
@@ -3,8 +3,8 @@ markdown: kramdown
 
 # These allow the documentation to be updated with nerw releases
 # of Spark, Scala, and Mesos.
-SPARK_VERSION: 1.0.2
-SPARK_VERSION_SHORT: 1.0.2
+SPARK_VERSION: 1.0.3
+SPARK_VERSION_SHORT: 1.0.3
 SCALA_BINARY_VERSION: 2.10
 SCALA_VERSION: 2.10.4
 MESOS_VERSION: 0.18.1

http://git-wip-us.apache.org/repos/asf/spark/blob/3fbe3447/extras/java8-tests/pom.xml
--
diff --git a/extras/java8-tests/pom.xml b/extras/java8-tests/pom.xml
index b4ad113..4944d3d 100644
--- a/extras/java8-tests/pom.xml
+++ b/extras/java8-tests/pom.xml
@@ -20,7 +20,7 @@
   parent
 groupIdorg.apache.spark/groupId
 artifactIdspark-parent/artifactId
-version1.0.2-SNAPSHOT/version
+version1.0.3-SNAPSHOT/version
 relativePath../../pom.xml/relativePath
   /parent
 

http://git-wip-us.apache.org/repos/asf/spark/blob/3fbe3447/project/SparkBuild.scala
--
diff --git a/project/SparkBuild.scala b/project/SparkBuild.scala
index 5b687e7..5328ad3 100644
--- a/project/SparkBuild.scala
+++ b/project/SparkBuild.scala
@@ -32,7 +32,7 @@ import scala.collection.JavaConversions._
 // import com.jsuereth.pgp.sbtplugin.PgpKeys._
 
 object SparkBuild extends Build {
-  val SPARK_VERSION = 1.0.2
+  val SPARK_VERSION = 1.0.3-SNAPSHOT
   val SPARK_VERSION_SHORT = SPARK_VERSION.replaceAll(-SNAPSHOT, )
 
   // Hadoop version to build against. For example, 1.0.4 for Apache 
releases, or

http://git-wip-us.apache.org/repos/asf/spark/blob/3fbe3447/project/project/SparkPluginBuild.scala
--
diff --git a/project/project/SparkPluginBuild.scala 
b/project/project/SparkPluginBuild.scala
index 722d26f..707c9d0 100644
--- a/project/project/SparkPluginBuild.scala
+++ b/project/project/SparkPluginBuild.scala
@@ -26,7 +26,7 @@ import sbt.Keys._
 object SparkPluginDef extends Build {
   lazy val root = Project(plugins, file(.)) dependsOn(sparkStyle)
   lazy val sparkStyle = Project(spark-style, file(spark-style), settings = 
styleSettings)
-  val sparkVersion = 1.0.2
+  val sparkVersion = 1.0.3-SNAPSHOT
   // There is actually no need to publish this artifact.
   def styleSettings = Defaults.defaultSettings ++ Seq (
 name :=  spark-style,


svn commit: r6047 - /dev/spark/spark-1.0.2-rc1/ /release/spark/spark-1.0.2/

2014-08-04 Thread tdas
Author: tdas
Date: Tue Aug  5 01:53:43 2014
New Revision: 6047

Log:
Moving Spark 1.0.2 from dev to release.

Added:
release/spark/spark-1.0.2/
  - copied from r6046, dev/spark/spark-1.0.2-rc1/
Removed:
dev/spark/spark-1.0.2-rc1/


-
To unsubscribe, e-mail: commits-unsubscr...@spark.apache.org
For additional commands, e-mail: commits-h...@spark.apache.org



git commit: [SPARK-1022][Streaming] Add Kafka real unit test

2014-08-05 Thread tdas
Repository: spark
Updated Branches:
  refs/heads/master 184048f80 - e87075df9


[SPARK-1022][Streaming] Add Kafka real unit test

This PR is a updated version of (https://github.com/apache/spark/pull/557) to 
actually test sending and receiving data through Kafka, and fix previous flaky 
issues.

@tdas, would you mind reviewing this PR? Thanks a lot.

Author: jerryshao saisai.s...@intel.com

Closes #1751 from jerryshao/kafka-unit-test and squashes the following commits:

b6a505f [jerryshao] code refactor according to comments
5222330 [jerryshao] Change JavaKafkaStreamSuite to better test it
5525f10 [jerryshao] Fix flaky issue of Kafka real unit test
4559310 [jerryshao] Minor changes for Kafka unit test
860f649 [jerryshao] Minor style changes, and tests ignored due to flakiness
796d4ca [jerryshao] Add real Kafka streaming test


Project: http://git-wip-us.apache.org/repos/asf/spark/repo
Commit: http://git-wip-us.apache.org/repos/asf/spark/commit/e87075df
Tree: http://git-wip-us.apache.org/repos/asf/spark/tree/e87075df
Diff: http://git-wip-us.apache.org/repos/asf/spark/diff/e87075df

Branch: refs/heads/master
Commit: e87075df977a539e4a1684045a7bd66c36285174
Parents: 184048f
Author: jerryshao saisai.s...@intel.com
Authored: Tue Aug 5 10:40:28 2014 -0700
Committer: Tathagata Das tathagata.das1...@gmail.com
Committed: Tue Aug 5 10:40:28 2014 -0700

--
 external/kafka/pom.xml  |   6 +
 .../streaming/kafka/JavaKafkaStreamSuite.java   | 125 ++--
 .../streaming/kafka/KafkaStreamSuite.scala  | 197 +--
 3 files changed, 293 insertions(+), 35 deletions(-)
--


http://git-wip-us.apache.org/repos/asf/spark/blob/e87075df/external/kafka/pom.xml
--
diff --git a/external/kafka/pom.xml b/external/kafka/pom.xml
index daf0336..2aee999 100644
--- a/external/kafka/pom.xml
+++ b/external/kafka/pom.xml
@@ -71,6 +71,12 @@
   /exclusions
 /dependency
 dependency
+  groupIdnet.sf.jopt-simple/groupId
+  artifactIdjopt-simple/artifactId
+  version3.2/version
+  scopetest/scope
+/dependency
+dependency
   groupIdorg.scalatest/groupId
   artifactIdscalatest_${scala.binary.version}/artifactId
   scopetest/scope

http://git-wip-us.apache.org/repos/asf/spark/blob/e87075df/external/kafka/src/test/java/org/apache/spark/streaming/kafka/JavaKafkaStreamSuite.java
--
diff --git 
a/external/kafka/src/test/java/org/apache/spark/streaming/kafka/JavaKafkaStreamSuite.java
 
b/external/kafka/src/test/java/org/apache/spark/streaming/kafka/JavaKafkaStreamSuite.java
index 9f8046b..0571454 100644
--- 
a/external/kafka/src/test/java/org/apache/spark/streaming/kafka/JavaKafkaStreamSuite.java
+++ 
b/external/kafka/src/test/java/org/apache/spark/streaming/kafka/JavaKafkaStreamSuite.java
@@ -17,31 +17,118 @@
 
 package org.apache.spark.streaming.kafka;
 
+import java.io.Serializable;
 import java.util.HashMap;
+import java.util.List;
+
+import scala.Predef;
+import scala.Tuple2;
+import scala.collection.JavaConverters;
+
+import junit.framework.Assert;
 
-import org.apache.spark.streaming.api.java.JavaPairReceiverInputDStream;
-import org.junit.Test;
-import com.google.common.collect.Maps;
 import kafka.serializer.StringDecoder;
+
+import org.apache.spark.api.java.JavaPairRDD;
+import org.apache.spark.api.java.function.Function;
 import org.apache.spark.storage.StorageLevel;
+import org.apache.spark.streaming.Duration;
 import org.apache.spark.streaming.LocalJavaStreamingContext;
+import org.apache.spark.streaming.api.java.JavaDStream;
+import org.apache.spark.streaming.api.java.JavaPairDStream;
+import org.apache.spark.streaming.api.java.JavaStreamingContext;
+
+import org.junit.Test;
+import org.junit.After;
+import org.junit.Before;
+
+public class JavaKafkaStreamSuite extends LocalJavaStreamingContext implements 
Serializable {
+  private transient KafkaStreamSuite testSuite = new KafkaStreamSuite();
+
+  @Before
+  @Override
+  public void setUp() {
+testSuite.beforeFunction();
+System.clearProperty(spark.driver.port);
+//System.setProperty(spark.streaming.clock, 
org.apache.spark.streaming.util.SystemClock);
+ssc = new JavaStreamingContext(local[2], test, new Duration(1000));
+  }
+
+  @After
+  @Override
+  public void tearDown() {
+ssc.stop();
+ssc = null;
+System.clearProperty(spark.driver.port);
+testSuite.afterFunction();
+  }
 
-public class JavaKafkaStreamSuite extends LocalJavaStreamingContext {
   @Test
-  public void testKafkaStream() {
-HashMapString, Integer topics = Maps.newHashMap();
-
-// tests the API, does not actually test data receiving
-JavaPairReceiverInputDStreamString, String test1 =
-KafkaUtils.createStream(ssc

git commit: [SPARK-1022][Streaming] Add Kafka real unit test

2014-08-05 Thread tdas
Repository: spark
Updated Branches:
  refs/heads/branch-1.1 075ba6781 - b92a45058


[SPARK-1022][Streaming] Add Kafka real unit test

This PR is a updated version of (https://github.com/apache/spark/pull/557) to 
actually test sending and receiving data through Kafka, and fix previous flaky 
issues.

@tdas, would you mind reviewing this PR? Thanks a lot.

Author: jerryshao saisai.s...@intel.com

Closes #1751 from jerryshao/kafka-unit-test and squashes the following commits:

b6a505f [jerryshao] code refactor according to comments
5222330 [jerryshao] Change JavaKafkaStreamSuite to better test it
5525f10 [jerryshao] Fix flaky issue of Kafka real unit test
4559310 [jerryshao] Minor changes for Kafka unit test
860f649 [jerryshao] Minor style changes, and tests ignored due to flakiness
796d4ca [jerryshao] Add real Kafka streaming test


Project: http://git-wip-us.apache.org/repos/asf/spark/repo
Commit: http://git-wip-us.apache.org/repos/asf/spark/commit/b92a4505
Tree: http://git-wip-us.apache.org/repos/asf/spark/tree/b92a4505
Diff: http://git-wip-us.apache.org/repos/asf/spark/diff/b92a4505

Branch: refs/heads/branch-1.1
Commit: b92a450583989470ff53b62c124d908ad661e29a
Parents: 075ba67
Author: jerryshao saisai.s...@intel.com
Authored: Tue Aug 5 10:40:28 2014 -0700
Committer: Tathagata Das tathagata.das1...@gmail.com
Committed: Tue Aug 5 10:43:35 2014 -0700

--
 external/kafka/pom.xml  |   6 +
 .../streaming/kafka/JavaKafkaStreamSuite.java   | 125 ++--
 .../streaming/kafka/KafkaStreamSuite.scala  | 197 +--
 3 files changed, 293 insertions(+), 35 deletions(-)
--


http://git-wip-us.apache.org/repos/asf/spark/blob/b92a4505/external/kafka/pom.xml
--
diff --git a/external/kafka/pom.xml b/external/kafka/pom.xml
index daf0336..2aee999 100644
--- a/external/kafka/pom.xml
+++ b/external/kafka/pom.xml
@@ -71,6 +71,12 @@
   /exclusions
 /dependency
 dependency
+  groupIdnet.sf.jopt-simple/groupId
+  artifactIdjopt-simple/artifactId
+  version3.2/version
+  scopetest/scope
+/dependency
+dependency
   groupIdorg.scalatest/groupId
   artifactIdscalatest_${scala.binary.version}/artifactId
   scopetest/scope

http://git-wip-us.apache.org/repos/asf/spark/blob/b92a4505/external/kafka/src/test/java/org/apache/spark/streaming/kafka/JavaKafkaStreamSuite.java
--
diff --git 
a/external/kafka/src/test/java/org/apache/spark/streaming/kafka/JavaKafkaStreamSuite.java
 
b/external/kafka/src/test/java/org/apache/spark/streaming/kafka/JavaKafkaStreamSuite.java
index 9f8046b..0571454 100644
--- 
a/external/kafka/src/test/java/org/apache/spark/streaming/kafka/JavaKafkaStreamSuite.java
+++ 
b/external/kafka/src/test/java/org/apache/spark/streaming/kafka/JavaKafkaStreamSuite.java
@@ -17,31 +17,118 @@
 
 package org.apache.spark.streaming.kafka;
 
+import java.io.Serializable;
 import java.util.HashMap;
+import java.util.List;
+
+import scala.Predef;
+import scala.Tuple2;
+import scala.collection.JavaConverters;
+
+import junit.framework.Assert;
 
-import org.apache.spark.streaming.api.java.JavaPairReceiverInputDStream;
-import org.junit.Test;
-import com.google.common.collect.Maps;
 import kafka.serializer.StringDecoder;
+
+import org.apache.spark.api.java.JavaPairRDD;
+import org.apache.spark.api.java.function.Function;
 import org.apache.spark.storage.StorageLevel;
+import org.apache.spark.streaming.Duration;
 import org.apache.spark.streaming.LocalJavaStreamingContext;
+import org.apache.spark.streaming.api.java.JavaDStream;
+import org.apache.spark.streaming.api.java.JavaPairDStream;
+import org.apache.spark.streaming.api.java.JavaStreamingContext;
+
+import org.junit.Test;
+import org.junit.After;
+import org.junit.Before;
+
+public class JavaKafkaStreamSuite extends LocalJavaStreamingContext implements 
Serializable {
+  private transient KafkaStreamSuite testSuite = new KafkaStreamSuite();
+
+  @Before
+  @Override
+  public void setUp() {
+testSuite.beforeFunction();
+System.clearProperty(spark.driver.port);
+//System.setProperty(spark.streaming.clock, 
org.apache.spark.streaming.util.SystemClock);
+ssc = new JavaStreamingContext(local[2], test, new Duration(1000));
+  }
+
+  @After
+  @Override
+  public void tearDown() {
+ssc.stop();
+ssc = null;
+System.clearProperty(spark.driver.port);
+testSuite.afterFunction();
+  }
 
-public class JavaKafkaStreamSuite extends LocalJavaStreamingContext {
   @Test
-  public void testKafkaStream() {
-HashMapString, Integer topics = Maps.newHashMap();
-
-// tests the API, does not actually test data receiving
-JavaPairReceiverInputDStreamString, String test1

svn commit: r1616049 - in /spark: ./ _layouts/ news/_posts/ releases/_posts/ site/ site/docs/ site/docs/1.0.2/ site/docs/1.0.2/api/ site/docs/1.0.2/api/java/ site/docs/1.0.2/api/java/org/ site/docs/1.

2014-08-05 Thread tdas
Author: tdas
Date: Tue Aug  5 23:40:54 2014
New Revision: 1616049

URL: http://svn.apache.org/r1616049
Log:
Adding Spark 1.0.2


[This commit notification would consist of 386 parts, 
which exceeds the limit of 50 ones, so it was shortened to the summary.]

-
To unsubscribe, e-mail: commits-unsubscr...@spark.apache.org
For additional commands, e-mail: commits-h...@spark.apache.org



svn commit: r1616050 - in /spark: news/_posts/2014-08-05-spark-1-0-2-released.md site/news/index.html site/news/spark-1-0-2-released.html

2014-08-05 Thread tdas
Author: tdas
Date: Tue Aug  5 23:43:56 2014
New Revision: 1616050

URL: http://svn.apache.org/r1616050
Log:
Small update to Spark 1.0.2.

Modified:
spark/news/_posts/2014-08-05-spark-1-0-2-released.md
spark/site/news/index.html
spark/site/news/spark-1-0-2-released.html

Modified: spark/news/_posts/2014-08-05-spark-1-0-2-released.md
URL: 
http://svn.apache.org/viewvc/spark/news/_posts/2014-08-05-spark-1-0-2-released.md?rev=1616050r1=1616049r2=1616050view=diff
==
--- spark/news/_posts/2014-08-05-spark-1-0-2-released.md (original)
+++ spark/news/_posts/2014-08-05-spark-1-0-2-released.md Tue Aug  5 23:43:56 
2014
@@ -11,6 +11,6 @@ meta:
   _edit_last: '4'
   _wpas_done_all: '1'
 ---
-We are happy to announce the availability of a 
href={{site.url}}releases/spark-release-1-0-2.html title=Spark Release 
1.0.2Spark 1.0.2/a! This release includes contributions from 30 developers. 
Spark 1.0.2 includes fixes across several areas of Spark, including the core 
API, PySpark, and MLlib.
+We are happy to announce the availability of a 
href={{site.url}}releases/spark-release-1-0-2.html title=Spark Release 
1.0.2Spark 1.0.2/a! This release includes contributions from 30 developers. 
Spark 1.0.2 includes fixes across several areas of Spark, including the core 
API, Streaming, PySpark, and MLlib.
 
 Visit the a href={{site.url}}releases/spark-release-1-0-2.html title=Spark 
Release 1.0.2release notes/a to read about this release or a 
href={{site.url}}downloads.htmldownload/a the release today.

Modified: spark/site/news/index.html
URL: 
http://svn.apache.org/viewvc/spark/site/news/index.html?rev=1616050r1=1616049r2=1616050view=diff
==
--- spark/site/news/index.html (original)
+++ spark/site/news/index.html Tue Aug  5 23:43:56 2014
@@ -164,7 +164,7 @@
   h3 class=entry-titlea href=/news/spark-1-0-2-released.htmlSpark 
1.0.2 released/a/h3
   div class=entry-dateAugust 5, 2014/div
 /header
-div class=entry-contentpWe are happy to announce the availability of 
a href=/releases/spark-release-1-0-2.html title=Spark Release 1.0.2Spark 
1.0.2/a! This release includes contributions from 30 developers. Spark 1.0.2 
includes fixes across several areas of Spark, including the core API, PySpark, 
and MLlib./p
+div class=entry-contentpWe are happy to announce the availability of 
a href=/releases/spark-release-1-0-2.html title=Spark Release 1.0.2Spark 
1.0.2/a! This release includes contributions from 30 developers. Spark 1.0.2 
includes fixes across several areas of Spark, including the core API, 
Streaming, PySpark, and MLlib./p
 
 /div
   /article

Modified: spark/site/news/spark-1-0-2-released.html
URL: 
http://svn.apache.org/viewvc/spark/site/news/spark-1-0-2-released.html?rev=1616050r1=1616049r2=1616050view=diff
==
--- spark/site/news/spark-1-0-2-released.html (original)
+++ spark/site/news/spark-1-0-2-released.html Tue Aug  5 23:43:56 2014
@@ -160,7 +160,7 @@
 h2Spark 1.0.2 released/h2
 
 
-pWe are happy to announce the availability of a 
href=/releases/spark-release-1-0-2.html title=Spark Release 1.0.2Spark 
1.0.2/a! This release includes contributions from 30 developers. Spark 1.0.2 
includes fixes across several areas of Spark, including the core API, PySpark, 
and MLlib./p
+pWe are happy to announce the availability of a 
href=/releases/spark-release-1-0-2.html title=Spark Release 1.0.2Spark 
1.0.2/a! This release includes contributions from 30 developers. Spark 1.0.2 
includes fixes across several areas of Spark, including the core API, 
Streaming, PySpark, and MLlib./p
 
 pVisit the a href=/releases/spark-release-1-0-2.html title=Spark Release 
1.0.2release notes/a to read about this release or a 
href=/downloads.htmldownload/a the release today./p
 



-
To unsubscribe, e-mail: commits-unsubscr...@spark.apache.org
For additional commands, e-mail: commits-h...@spark.apache.org



  1   2   3   4   5   6   7   8   9   >