Hi Pat and others, I see that cf/CooccuranceAnalysis.scala is currently under spark. Is there a specific reason? I see that the code itself is completely spark agnostic. I tried moving the code under math-scala/src/main/scala/org/apache/mahout/math/cf/ with the following trivial patch:
diff --git a/spark/src/main/scala/org/apache/mahout/cf/CooccurrenceAnalysis.scala b/spark/src/main/scala/org/apache/mahout/cf/CooccurrenceAnalysis.scala index ee44f90..bd20956 100644 --- a/spark/src/main/scala/org/apache/mahout/cf/CooccurrenceAnalysis.scala +++ b/spark/src/main/scala/org/apache/mahout/cf/CooccurrenceAnalysis.scala @@ -22,7 +22,6 @@ import scalabindings._ import RLikeOps._ import drm._ import RLikeDrmOps._ -import org.apache.mahout.sparkbindings._ import scala.collection.JavaConversions._ import org.apache.mahout.math.stats.LogLikelihood and it seems to work just fine. From what I see, this should work just fine on H2O as well with no changes.. Why give up generality and make it spark specific? Thanks
