Repository: mahout Updated Branches: refs/heads/master e7f247d8e -> 56a2305ed
(NOJIRA) minor test style Project: http://git-wip-us.apache.org/repos/asf/mahout/repo Commit: http://git-wip-us.apache.org/repos/asf/mahout/commit/56a2305e Tree: http://git-wip-us.apache.org/repos/asf/mahout/tree/56a2305e Diff: http://git-wip-us.apache.org/repos/asf/mahout/diff/56a2305e Branch: refs/heads/master Commit: 56a2305ed14a3f4706b10c394a27d3429c2103b4 Parents: e7f247d Author: Dmitriy Lyubimov <[email protected]> Authored: Thu Jul 31 13:28:17 2014 -0700 Committer: Dmitriy Lyubimov <[email protected]> Committed: Thu Jul 31 13:28:17 2014 -0700 ---------------------------------------------------------------------- .../DistributedDecompositionsSuiteBase.scala | 6 +-- .../io/MahoutKryoRegistrator.scala | 8 +--- .../mahout/sparkbindings/blas/ABtSuite.scala | 8 ++-- .../mahout/sparkbindings/blas/AewBSuite.scala | 40 ++++++++++---------- .../mahout/sparkbindings/blas/AtASuite.scala | 1 - .../mahout/sparkbindings/blas/AtSuite.scala | 9 ++--- 6 files changed, 32 insertions(+), 40 deletions(-) ---------------------------------------------------------------------- http://git-wip-us.apache.org/repos/asf/mahout/blob/56a2305e/math-scala/src/test/scala/org/apache/mahout/math/decompositions/DistributedDecompositionsSuiteBase.scala ---------------------------------------------------------------------- diff --git a/math-scala/src/test/scala/org/apache/mahout/math/decompositions/DistributedDecompositionsSuiteBase.scala b/math-scala/src/test/scala/org/apache/mahout/math/decompositions/DistributedDecompositionsSuiteBase.scala index 1aa3a20..d37ab17 100644 --- a/math-scala/src/test/scala/org/apache/mahout/math/decompositions/DistributedDecompositionsSuiteBase.scala +++ b/math-scala/src/test/scala/org/apache/mahout/math/decompositions/DistributedDecompositionsSuiteBase.scala @@ -47,11 +47,11 @@ trait DistributedDecompositionsSuiteBase extends DistributedMahoutSuite with Mat (8, 6, 7, 8) ) - val A = drmParallelize(inCoreA, numPartitions = 2) - val (drmQ, inCoreR) = dqrThin(A, checkRankDeficiency = false) + val drmA = drmParallelize(inCoreA, numPartitions = 2) + val (drmQ, inCoreR) = dqrThin(drmA, checkRankDeficiency = false) // Assert optimizer still knows Q and A are identically partitioned - drmQ.partitioningTag should equal(A.partitioningTag) + drmQ.partitioningTag should equal(drmA.partitioningTag) // drmQ.rdd.partitions.size should be(A.rdd.partitions.size) // http://git-wip-us.apache.org/repos/asf/mahout/blob/56a2305e/spark/src/main/scala/org/apache/mahout/sparkbindings/io/MahoutKryoRegistrator.scala ---------------------------------------------------------------------- diff --git a/spark/src/main/scala/org/apache/mahout/sparkbindings/io/MahoutKryoRegistrator.scala b/spark/src/main/scala/org/apache/mahout/sparkbindings/io/MahoutKryoRegistrator.scala index 79c7585..22e31cc 100644 --- a/spark/src/main/scala/org/apache/mahout/sparkbindings/io/MahoutKryoRegistrator.scala +++ b/spark/src/main/scala/org/apache/mahout/sparkbindings/io/MahoutKryoRegistrator.scala @@ -24,10 +24,7 @@ import org.apache.mahout.math._ import org.apache.spark.serializer.KryoRegistrator import org.apache.mahout.sparkbindings._ - -/** - * - */ +/** Kryo serialization registrator for Mahout */ class MahoutKryoRegistrator extends KryoRegistrator { override def registerClasses(kryo: Kryo) = { @@ -38,6 +35,3 @@ class MahoutKryoRegistrator extends KryoRegistrator { kryo.register(classOf[com.google.common.collect.HashBiMap[String, Int]], new JavaSerializer()); } } - -object MahoutKryoRegistrator { -} http://git-wip-us.apache.org/repos/asf/mahout/blob/56a2305e/spark/src/test/scala/org/apache/mahout/sparkbindings/blas/ABtSuite.scala ---------------------------------------------------------------------- diff --git a/spark/src/test/scala/org/apache/mahout/sparkbindings/blas/ABtSuite.scala b/spark/src/test/scala/org/apache/mahout/sparkbindings/blas/ABtSuite.scala index 69dbcbf..12c9034 100644 --- a/spark/src/test/scala/org/apache/mahout/sparkbindings/blas/ABtSuite.scala +++ b/spark/src/test/scala/org/apache/mahout/sparkbindings/blas/ABtSuite.scala @@ -33,12 +33,12 @@ class ABtSuite extends FunSuite with DistributedSparkSuite { test("ABt") { val inCoreA = dense((1, 2, 3), (2, 3, 4), (3, 4, 5)) val inCoreB = dense((3, 4, 5), (5, 6, 7)) - val A = drmParallelize(m = inCoreA, numPartitions = 3) - val B = drmParallelize(m = inCoreB, numPartitions = 2) + val drmA = drmParallelize(m = inCoreA, numPartitions = 3) + val drmB = drmParallelize(m = inCoreB, numPartitions = 2) - val op = new OpABt(A, B) + val op = new OpABt(drmA, drmB) - val drm = new CheckpointedDrmSpark(ABt.abt(op, srcA = A, srcB = B), op.nrow, op.ncol) + val drm = new CheckpointedDrmSpark(ABt.abt(op, srcA = drmA, srcB = drmB), op.nrow, op.ncol) printf("AB' num partitions = %d.\n", drm.rdd.partitions.size) http://git-wip-us.apache.org/repos/asf/mahout/blob/56a2305e/spark/src/test/scala/org/apache/mahout/sparkbindings/blas/AewBSuite.scala ---------------------------------------------------------------------- diff --git a/spark/src/test/scala/org/apache/mahout/sparkbindings/blas/AewBSuite.scala b/spark/src/test/scala/org/apache/mahout/sparkbindings/blas/AewBSuite.scala index 661e2fe..be65e32 100644 --- a/spark/src/test/scala/org/apache/mahout/sparkbindings/blas/AewBSuite.scala +++ b/spark/src/test/scala/org/apache/mahout/sparkbindings/blas/AewBSuite.scala @@ -33,14 +33,14 @@ class AewBSuite extends FunSuite with DistributedSparkSuite { test("A * B Hadamard") { val inCoreA = dense((1, 2, 3), (2, 3, 4), (3, 4, 5), (7, 8, 9)) val inCoreB = dense((3, 4, 5), (5, 6, 7), (0, 0, 0), (9, 8, 7)) - val A = drmParallelize(m = inCoreA, numPartitions = 2) - val B = drmParallelize(m = inCoreB) + val drmA = drmParallelize(m = inCoreA, numPartitions = 2) + val drmB = drmParallelize(m = inCoreB) - val op = new OpAewB(A, B, "*") + val op = new OpAewB(drmA, drmB, "*") - val M = new CheckpointedDrmSpark(AewB.a_ew_b(op, srcA = A, srcB = B), op.nrow, op.ncol) + val drmM = new CheckpointedDrmSpark(AewB.a_ew_b(op, srcA = drmA, srcB = drmB), op.nrow, op.ncol) - val inCoreM = M.collect + val inCoreM = drmM.collect val inCoreMControl = inCoreA * inCoreB assert((inCoreM - inCoreMControl).norm < 1E-10) @@ -50,14 +50,14 @@ class AewBSuite extends FunSuite with DistributedSparkSuite { test("A + B Elementwise") { val inCoreA = dense((1, 2, 3), (2, 3, 4), (3, 4, 5), (7, 8, 9)) val inCoreB = dense((3, 4, 5), (5, 6, 7), (0, 0, 0), (9, 8, 7)) - val A = drmParallelize(m = inCoreA, numPartitions = 2) - val B = drmParallelize(m = inCoreB) + val drmA = drmParallelize(m = inCoreA, numPartitions = 2) + val drmB = drmParallelize(m = inCoreB) - val op = new OpAewB(A, B, "+") + val op = new OpAewB(drmA, drmB, "+") - val M = new CheckpointedDrmSpark(AewB.a_ew_b(op, srcA = A, srcB = B), op.nrow, op.ncol) + val drmM = new CheckpointedDrmSpark(AewB.a_ew_b(op, srcA = drmA, srcB = drmB), op.nrow, op.ncol) - val inCoreM = M.collect + val inCoreM = drmM.collect val inCoreMControl = inCoreA + inCoreB assert((inCoreM - inCoreMControl).norm < 1E-10) @@ -67,14 +67,14 @@ class AewBSuite extends FunSuite with DistributedSparkSuite { test("A - B Elementwise") { val inCoreA = dense((1, 2, 3), (2, 3, 4), (3, 4, 5), (7, 8, 9)) val inCoreB = dense((3, 4, 5), (5, 6, 7), (0, 0, 0), (9, 8, 7)) - val A = drmParallelize(m = inCoreA, numPartitions = 2) - val B = drmParallelize(m = inCoreB) + val drmA = drmParallelize(m = inCoreA, numPartitions = 2) + val drmB = drmParallelize(m = inCoreB) - val op = new OpAewB(A, B, "-") + val op = new OpAewB(drmA, drmB, "-") - val M = new CheckpointedDrmSpark(AewB.a_ew_b(op, srcA = A, srcB = B), op.nrow, op.ncol) + val drmM = new CheckpointedDrmSpark(AewB.a_ew_b(op, srcA = drmA, srcB = drmB), op.nrow, op.ncol) - val inCoreM = M.collect + val inCoreM = drmM.collect val inCoreMControl = inCoreA - inCoreB assert((inCoreM - inCoreMControl).norm < 1E-10) @@ -84,14 +84,14 @@ class AewBSuite extends FunSuite with DistributedSparkSuite { test("A / B Elementwise") { val inCoreA = dense((1, 2, 3), (2, 3, 4), (3, 4, 0), (7, 8, 9)) val inCoreB = dense((3, 4, 5), (5, 6, 7), (10, 20, 30), (9, 8, 7)) - val A = drmParallelize(m = inCoreA, numPartitions = 2) - val B = drmParallelize(m = inCoreB) + val drmA = drmParallelize(m = inCoreA, numPartitions = 2) + val drmB = drmParallelize(m = inCoreB) - val op = new OpAewB(A, B, "/") + val op = new OpAewB(drmA, drmB, "/") - val M = new CheckpointedDrmSpark(AewB.a_ew_b(op, srcA = A, srcB = B), op.nrow, op.ncol) + val drmM = new CheckpointedDrmSpark(AewB.a_ew_b(op, srcA = drmA, srcB = drmB), op.nrow, op.ncol) - val inCoreM = M.collect + val inCoreM = drmM.collect val inCoreMControl = inCoreA / inCoreB assert((inCoreM - inCoreMControl).norm < 1E-10) http://git-wip-us.apache.org/repos/asf/mahout/blob/56a2305e/spark/src/test/scala/org/apache/mahout/sparkbindings/blas/AtASuite.scala ---------------------------------------------------------------------- diff --git a/spark/src/test/scala/org/apache/mahout/sparkbindings/blas/AtASuite.scala b/spark/src/test/scala/org/apache/mahout/sparkbindings/blas/AtASuite.scala index 49b3f46..c31f27c 100644 --- a/spark/src/test/scala/org/apache/mahout/sparkbindings/blas/AtASuite.scala +++ b/spark/src/test/scala/org/apache/mahout/sparkbindings/blas/AtASuite.scala @@ -45,5 +45,4 @@ class AtASuite extends FunSuite with DistributedSparkSuite { } - } http://git-wip-us.apache.org/repos/asf/mahout/blob/56a2305e/spark/src/test/scala/org/apache/mahout/sparkbindings/blas/AtSuite.scala ---------------------------------------------------------------------- diff --git a/spark/src/test/scala/org/apache/mahout/sparkbindings/blas/AtSuite.scala b/spark/src/test/scala/org/apache/mahout/sparkbindings/blas/AtSuite.scala index 0123b78..16632ec 100644 --- a/spark/src/test/scala/org/apache/mahout/sparkbindings/blas/AtSuite.scala +++ b/spark/src/test/scala/org/apache/mahout/sparkbindings/blas/AtSuite.scala @@ -30,16 +30,15 @@ class AtSuite extends FunSuite with DistributedSparkSuite { test("At") { val inCoreA = dense((1, 2, 3), (2, 3, 4), (3, 4, 5)) - val A = drmParallelize(m = inCoreA, numPartitions = 2) + val drmA = drmParallelize(m = inCoreA, numPartitions = 2) - val op = new OpAt(A) - val AtDrm = new CheckpointedDrmSpark(rdd = At.at(op, srcA = A), _nrow = op.nrow, _ncol = op.ncol) - val inCoreAt = AtDrm.collect + val op = new OpAt(drmA) + val drmAt = new CheckpointedDrmSpark(rdd = At.at(op, srcA = drmA), _nrow = op.nrow, _ncol = op.ncol) + val inCoreAt = drmAt.collect val inCoreControlAt = inCoreA.t println(inCoreAt) assert((inCoreAt - inCoreControlAt).norm < 1E-5) - } }
