Hi friends,
I have a code which is written in Scala. The scala version 2.10.4 and Spark version 1.5.2 are used to run the code.
I would like to upgrade the code to the most updated version of spark, meaning 2.1.0.
Here is the build.sbt:
import AssemblyKeys._
assemblySettings
name := "proxcocoa"
version := "0.1"
scalaVersion := "2.10.4"
parallelExecution in Test := false
{
val excludeHadoop = ExclusionRule(organization = "org.apache.hadoop")
libraryDependencies ++= Seq(
"org.slf4j" % "slf4j-api" % "1.7.2",
"org.slf4j" % "slf4j-log4j12" % "1.7.2",
"org.scalatest" %% "scalatest" % "1.9.1" % "test",
"org.apache.spark" % "spark-core_2.10" % "1.5.2" excludeAll(excludeHadoop),
"org.apache.spark" % "spark-mllib_2.10" % "1.5.2" excludeAll(excludeHadoop),
"org.apache.spark" % "spark-sql_2.10" % "1.5.2" excludeAll(excludeHadoop),
"org.apache.commons" % "commons-compress" % "1.7",
"commons-io" % "commons-io" % "2.4",
"org.scalanlp" % "breeze_2.10" % "0.11.2",
"com.github.fommil.netlib" % "all" % "1.1.2" pomOnly(),
"com.github.scopt" %% "scopt" % "3.3.0"
)
}
{
val defaultHadoopVersion = "1.0.4"
val hadoopVersion =
scala.util.Properties.envOrElse("SPARK_HADOOP_VERSION", defaultHadoopVersion)
libraryDependencies += "org.apache.hadoop" % "hadoop-client" % hadoopVersion
}
libraryDependencies += "org.apache.spark" % "spark-streaming_2.10" % "1.5.0"
resolvers ++= Seq(
"Local Maven Repository" at Path.userHome.asFile.toURI.toURL + ".m2/repository",
"Typesafe" at "
http://repo.typesafe.com/typesafe/releases",
"Spray" at "
http://repo.spray.cc"
)
mergeStrategy in assembly <<= (mergeStrategy in assembly) { (old) =>
{
case PathList("javax", "servlet", xs @ _*) => MergeStrategy.first
case PathList(ps @ _*) if ps.last endsWith ".html" => MergeStrategy.first
case "application.conf" => MergeStrategy.concat
case "reference.conf" => MergeStrategy.concat
case "log4j.properties" => MergeStrategy.discard
case m if m.toLowerCase.endsWith("manifest.mf") => MergeStrategy.discard
case m if m.toLowerCase.matches("meta-inf.*\\.sf$") => MergeStrategy.discard
case _ => MergeStrategy.first
}
}
test in assembly := {}
-----------------------------------------------------------
I downloaded the spark 2.1.0 and change the version of spark and scalaversion in the build.sbt. But unfortunately, I was failed to run the code.