Rename SBT target to assemble-deps.

Project: http://git-wip-us.apache.org/repos/asf/incubator-spark/repo
Commit: http://git-wip-us.apache.org/repos/asf/incubator-spark/commit/0a4b76fc
Tree: http://git-wip-us.apache.org/repos/asf/incubator-spark/tree/0a4b76fc
Diff: http://git-wip-us.apache.org/repos/asf/incubator-spark/diff/0a4b76fc

Branch: refs/heads/master
Commit: 0a4b76fcc22568dd62e6be0301e3bda916bc6bc8
Parents: 1dcded4
Author: Shivaram Venkataraman <shiva...@eecs.berkeley.edu>
Authored: Wed Oct 16 17:05:46 2013 -0700
Committer: Shivaram Venkataraman <shiva...@eecs.berkeley.edu>
Committed: Wed Oct 16 17:05:46 2013 -0700

----------------------------------------------------------------------
 project/SparkBuild.scala | 10 +++++-----
 1 file changed, 5 insertions(+), 5 deletions(-)
----------------------------------------------------------------------


http://git-wip-us.apache.org/repos/asf/incubator-spark/blob/0a4b76fc/project/SparkBuild.scala
----------------------------------------------------------------------
diff --git a/project/SparkBuild.scala b/project/SparkBuild.scala
index d7e8207..dd26d7e 100644
--- a/project/SparkBuild.scala
+++ b/project/SparkBuild.scala
@@ -60,7 +60,7 @@ object SparkBuild extends Build {
   lazy val assemblyProj = Project("assembly", file("assembly"), settings = 
assemblyProjSettings)
     .dependsOn(core, bagel, mllib, repl, streaming) dependsOn(maybeYarn: _*)
 
-  lazy val spark = TaskKey[Unit]("spark", "Build assembly of dependencies and 
spark packages")
+  lazy val assembleDeps = TaskKey[Unit]("assemble-deps", "Build assembly of 
dependencies and packages Spark projects")
 
   // A configuration to set an alternative publishLocalConfiguration
   lazy val MavenCompile = config("m2r") extend(Compile)
@@ -76,12 +76,12 @@ object SparkBuild extends Build {
   // Conditionally include the yarn sub-project
   lazy val maybeYarn = if(isYarnEnabled) Seq[ClasspathDependency](yarn) else 
Seq[ClasspathDependency]()
   lazy val maybeYarnRef = if(isYarnEnabled) Seq[ProjectReference](yarn) else 
Seq[ProjectReference]()
-  lazy val allProjects = Seq[ProjectReference](
-    core, repl, examples, bagel, streaming, mllib, tools, assemblyProj) ++ 
maybeYarnRef
 
-  // Everything except assembly and examples belongs to packageProjects
+  // Everything except assembly, tools and examples belong to packageProjects
   lazy val packageProjects = Seq[ProjectReference](core, repl, bagel, 
streaming, mllib) ++ maybeYarnRef
 
+  lazy val allProjects = packageProjects ++ Seq[ProjectReference](examples, 
tools, assemblyProj)
+
   def sharedSettings = Defaults.defaultSettings ++ Seq(
     organization := "org.apache.spark",
     version := "0.9.0-incubating-SNAPSHOT",
@@ -308,7 +308,7 @@ object SparkBuild extends Build {
 
   def assemblyProjSettings = sharedSettings ++ Seq(
     name := "spark-assembly",
-    spark in Compile <<= (packageProjects.map(packageBin in Compile in _) ++ 
Seq(packageDependency in Compile)).dependOn,
+    assembleDeps in Compile <<= (packageProjects.map(packageBin in Compile in 
_) ++ Seq(packageDependency in Compile)).dependOn,
     jarName in assembly <<= version map { v => "spark-assembly-" + v + 
"-hadoop" + hadoopVersion + ".jar" },
     jarName in packageDependency <<= version map { v => "spark-assembly-" + v 
+ "-hadoop" + hadoopVersion + "-deps.jar" }
   ) ++ assemblySettings ++ extraAssemblySettings

Reply via email to