I built Spark 1.2.0 succesfully, but was unable to build my Spark program under 1.2.0 with sbt assembly & my build.sbt file. It contains:
I tried: "org.apache.spark" %% "spark-sql" % "1.2.0", "org.apache.spark" %% "spark-core" % "1.2.0", and "org.apache.spark" %% "spark-sql" % "1.2.0-SNAPSHOT", "org.apache.spark" %% "spark-core" % "1.2.0-SNAPSHOT", but I get errors like: [warn] :::::::::::::::::::::::::::::::::::::::::::::: [warn] :: UNRESOLVED DEPENDENCIES :: [warn] :::::::::::::::::::::::::::::::::::::::::::::: [warn] :: org.apache.spark#spark-sql_2.10;1.2.0: not found [warn] :: org.apache.spark#spark-core_2.10;1.2.0: not found [warn] :::::::::::::::::::::::::::::::::::::::::::::: sbt.ResolveException: unresolved dependency: org.apache.spark#spark-sql_2.10;1.2.0: not found unresolved dependency: org.apache.spark#spark-core_2.10;1.2.0: not found ... [error] (*:update) sbt.ResolveException: unresolved dependency: org.apache.spark#spark-sql_2.10;1.2.0: not found [error] unresolved dependency: org.apache.spark#spark-core_2.10;1.2.0: not found Do I need to configure my build.sbt to point to my local spark 1.2.0 repository? How? Thanks, - Arun