Actually it has several Spark deps like having an SparkContext, SparkConf, and and rdd for file I/O Please look before you vote. I’ve been waving this flag for awhile—I/O is not engine neutral.
On Jun 19, 2014, at 11:41 AM, Sebastian Schelter <[email protected]> wrote: Hi Anand, Yes, this should not contain anything spark-specific. +1 for moving it. --sebastian On 06/19/2014 08:38 PM, Anand Avati wrote: > Hi Pat and others, > I see that cf/CooccuranceAnalysis.scala is currently under spark. Is there > a specific reason? I see that the code itself is completely spark agnostic. > I tried moving the code under > math-scala/src/main/scala/org/apache/mahout/math/cf/ with the following > trivial patch: > > diff --git > a/spark/src/main/scala/org/apache/mahout/cf/CooccurrenceAnalysis.scala > b/spark/src/main/scala/org/apache/mahout/cf/CooccurrenceAnalysis.scala > index ee44f90..bd20956 100644 > --- a/spark/src/main/scala/org/apache/mahout/cf/CooccurrenceAnalysis.scala > +++ b/spark/src/main/scala/org/apache/mahout/cf/CooccurrenceAnalysis.scala > @@ -22,7 +22,6 @@ import scalabindings._ > import RLikeOps._ > import drm._ > import RLikeDrmOps._ > -import org.apache.mahout.sparkbindings._ > import scala.collection.JavaConversions._ > import org.apache.mahout.math.stats.LogLikelihood > > > and it seems to work just fine. From what I see, this should work just fine > on H2O as well with no changes.. Why give up generality and make it spark > specific? > > Thanks >
