Greetings mahout users, I have been trying to use mahout samsara as a library with scala/spark, but I haven't been successful in doing so.
I am running spark 1.6.0 binaries, didn't build it myself. However, I tried both readily available binaries on Apache mirrors, and cloning and compiling mahout's repo, but neither worked. I keep getting Exception in thread "main" java.lang.NoClassDefFoundError: org/apache/mahout/sparkbindings/SparkDistributedContext The way I am doing things is: I have spark in ~/spark-1.6 and mahout in ~/mahout I have set both $SPARK_HOME and $MAHOUT_HOME accordingly, along with $MAHOUT_LOCAL=true and I have: ~/app1/build.sbt ~/app1/src/main/scala/App1.scala in build.sbt I have these lines to declare mahout dependecies: libraryDependencies += "org.apache.mahout" %% "mahout-math-scala" % "0.11.1" libraryDependencies += "org.apache.mahout" % "mahout-math" % "0.11.1" libraryDependencies += "org.apache.mahout" % "mahout-spark_2.10" % "0.11.1" along with other spark dependencies and in App1.scala, in the main function, I construct a context object using mahoutSparkContext, and of course, the sparkbindings are imported everything compiles successfully however, when I submit to spark, I get the above mentioned error. I have a general idea of why this is happening: because the compiled app1 jar depends on mahout-spark dependency jar but it cannot find it in the class path upon being submitted to spark. In the instructions I couldn't find how to explicitly add the mahout-spark dependency jar to the class path. The question is: Am I doing the configurations correctly or not? Sorry for the lengthy email Kind Regards, Bahaa