Running the example in the README gives a class not found:
"java.lang.NoClassDefFoundError:
it/unimi/dsi/fastutil/ints/Int2DoubleOpenHashMap"

If that's just us still using something that's been removed, it's not a
deal-breaker for me as long as we fix it in a quick point release.

Pending that being a simple fix my vote is +1 binding, and if Andy's not
back from vacation and his proxy works that's +2 binding from me and Andy.


bob $ export JAVA_HOME=/usr/lib/jvm/java-8-openjdk-amd64
bob $ export MAHOUT_HOME=//home/akm/a/src/test/
repository.apache.org/content/repositories/orgapachemahout-1052/org/apache/mahout/mahout/0.14.0
bob $ export SPARK_HOME=/home/akm/a/src/spark-2.1.0-bin-hadoop2.7
bob $ MASTER=local[2] mahout-0.14.0/bin/mahout spark-shell
Adding lib/ to CLASSPATH
Using Spark's default log4j profile:
org/apache/spark/log4j-defaults.properties
Setting default log level to "WARN".
To adjust logging level use sc.setLogLevel(newLevel). For SparkR, use
setLogLevel(newLevel).
19/03/04 09:07:44 WARN NativeCodeLoader: Unable to load native-hadoop
library for your platform... using builtin-java classes where applicable
19/03/04 09:07:44 WARN Utils: Your hostname, Bob resolves to a loopback
address: 127.0.1.1; using 10.0.1.2 instead (on interface eno1)
19/03/04 09:07:44 WARN Utils: Set SPARK_LOCAL_IP if you need to bind to
another address
19/03/04 09:07:53 WARN ObjectStore: Failed to get database global_temp,
returning NoSuchObjectException
Spark context Web UI available at http://10.0.1.2:4040
Spark context available as 'sc' (master = local[2], app id =
local-1551719265339).
Spark session available as 'spark'.
Loading /home/akm/a/src/test/
repository.apache.org/content/repositories/orgapachemahout-1052/org/apache/mahout/mahout/0.14.0/mahout-0.14.0/bin/load-shell.scala.
..
import org.apache.mahout.math._
import org.apache.mahout.math.scalabindings._
import org.apache.mahout.math.drm._
import org.apache.mahout.math.scalabindings.RLikeOps._
import org.apache.mahout.math.drm.RLikeDrmOps._
import org.apache.mahout.sparkbindings._
sdc: org.apache.mahout.sparkbindings.SparkDistributedContext =
org.apache.mahout.sparkbindings.SparkDistributedContext@749ffdc7

                _                 _
_ __ ___   __ _| |__   ___  _   _| |_
 '_ ` _ \ / _` | '_ \ / _ \| | | | __|
 | | | | (_| | | | | (_) | |_| | |_
_| |_| |_|\__,_|_| |_|\___/ \__,_|\__|  version 0.14.0



That file does not exist

Welcome to
      ____              __
     / __/__  ___ _____/ /__
    _\ \/ _ \/ _ `/ __/  '_/
   /___/ .__/\_,_/_/ /_/\_\   version 2.1.0
      /_/

Using Scala version 2.11.8 (OpenJDK 64-Bit Server VM, Java 1.8.0_191)
Type in expressions to have them evaluated.
Type :help for more information.

scala> :load /home/akm/a/src/test/
repository.apache.org/content/repositories/orgapachemahout-1052/org/apache/mahout/mahout/0.14.0/mahout-0.14.0/examples/bin/SparseSparseDrmTimer.mscala
Loading /home/akm/a/src/test/
repository.apache.org/content/repositories/orgapachemahout-1052/org/apache/mahout/mahout/0.14.0/mahout-0.14.0/examples/bin/SparseSparseDrmTimer.mscala.
..
timeSparseDRMMMul: (m: Int, n: Int, s: Int, para: Int, pctDense: Double,
seed: Long)Long

scala> timeSparseDRMMMul(1000,1000,1000,1,.02,1234L)
19/03/04 09:13:13 ERROR Executor: Exception in task 0.0 in stage 0.0 (TID 1)
java.lang.NoClassDefFoundError:
it/unimi/dsi/fastutil/ints/Int2DoubleOpenHashMap
    at
org.apache.mahout.math.RandomAccessSparseVector.<init>(RandomAccessSparseVector.java:49)
    at
org.apache.mahout.math.RandomAccessSparseVector.<init>(RandomAccessSparseVector.java:44)
    at
org.apache.mahout.sparkbindings.SparkEngine$$anonfun$11$$anonfun$apply$2.apply(SparkEngine.scala:200)
    at
org.apache.mahout.sparkbindings.SparkEngine$$anonfun$11$$anonfun$apply$2.apply(SparkEngine.scala:200)
    at
scala.collection.TraversableLike$$anonfun$map$1.apply(TraversableLike.scala:234)
    at
scala.collection.TraversableLike$$anonfun$map$1.apply(TraversableLike.scala:234)
    at scala.collection.immutable.Range.foreach(Range.scala:160)
    at scala.collection.TraversableLike$class.map(TraversableLike.scala:234)
    at scala.collection.AbstractTraversable.map(Traversable.scala:104)
    at
org.apache.mahout.sparkbindings.SparkEngine$$anonfun$11.apply(SparkEngine.scala:200)
    at
org.apache.mahout.sparkbindings.SparkEngine$$anonfun$11.apply(SparkEngine.scala:195)
    at scala.collection.Iterator$$anon$12.nextCur(Iterator.scala:434)
    at scala.collection.Iterator$$anon$12.hasNext(Iterator.scala:440)
    at scala.collection.Iterator$class.isEmpty(Iterator.scala:330)
    at scala.collection.AbstractIterator.isEmpty(Iterator.scala:1336)
    at
org.apache.mahout.sparkbindings.drm.package$$anonfun$blockify$1.apply(package.scala:55)
    at
org.apache.mahout.sparkbindings.drm.package$$anonfun$blockify$1.apply(package.scala:53)
    at
org.apache.spark.rdd.RDD$$anonfun$mapPartitions$1$$anonfun$apply$23.apply(RDD.scala:796)
    at
org.apache.spark.rdd.RDD$$anonfun$mapPartitions$1$$anonfun$apply$23.apply(RDD.scala:796)
    at
org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:38)
    at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:323)
    at org.apache.spark.rdd.RDD.iterator(RDD.scala:287)
    at
org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:38)
    at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:323)
    at org.apache.spark.rdd.RDD.iterator(RDD.scala:287)
    at
org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:38)
    at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:323)
    at org.apache.spark.rdd.RDD.iterator(RDD.scala:287)
    at
org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:96)
    at
org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:53)
    at org.apache.spark.scheduler.Task.run(Task.scala:99)
    at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:282)
    at
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
    at
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
    at java.lang.Thread.run(Thread.java:748)
19/03/04 09:13:13 ERROR Executor: Exception in task 0.0 in stage 1.0 (TID 0)
java.lang.NoClassDefFoundError:
it/unimi/dsi/fastutil/ints/Int2DoubleOpenHashMap
    at
org.apache.mahout.math.RandomAccessSparseVector.<init>(RandomAccessSparseVector.java:49)
    at
org.apache.mahout.math.RandomAccessSparseVector.<init>(RandomAccessSparseVector.java:44)
    at
org.apache.mahout.sparkbindings.SparkEngine$$anonfun$11$$anonfun$apply$2.apply(SparkEngine.scala:200)
    at
org.apache.mahout.sparkbindings.SparkEngine$$anonfun$11$$anonfun$apply$2.apply(SparkEngine.scala:200)
    at
scala.collection.TraversableLike$$anonfun$map$1.apply(TraversableLike.scala:234)
    at
scala.collection.TraversableLike$$anonfun$map$1.apply(TraversableLike.scala:234)
    at scala.collection.immutable.Range.foreach(Range.scala:160)
    at scala.collection.TraversableLike$class.map(TraversableLike.scala:234)
    at scala.collection.AbstractTraversable.map(Traversable.scala:104)
    at
org.apache.mahout.sparkbindings.SparkEngine$$anonfun$11.apply(SparkEngine.scala:200)
    at
org.apache.mahout.sparkbindings.SparkEngine$$anonfun$11.apply(SparkEngine.scala:195)
    at scala.collection.Iterator$$anon$12.nextCur(Iterator.scala:434)
    at scala.collection.Iterator$$anon$12.hasNext(Iterator.scala:440)
    at scala.collection.Iterator$class.isEmpty(Iterator.scala:330)
    at scala.collection.AbstractIterator.isEmpty(Iterator.scala:1336)
    at
org.apache.mahout.sparkbindings.drm.package$$anonfun$blockify$1.apply(package.scala:55)
    at
org.apache.mahout.sparkbindings.drm.package$$anonfun$blockify$1.apply(package.scala:53)
    at
org.apache.spark.rdd.RDD$$anonfun$mapPartitions$1$$anonfun$apply$23.apply(RDD.scala:796)
    at
org.apache.spark.rdd.RDD$$anonfun$mapPartitions$1$$anonfun$apply$23.apply(RDD.scala:796)
    at
org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:38)
    at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:323)
    at org.apache.spark.rdd.RDD.iterator(RDD.scala:287)
    at
org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:38)
    at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:323)
    at org.apache.spark.rdd.RDD.iterator(RDD.scala:287)
    at
org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:38)
    at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:323)
    at org.apache.spark.rdd.RDD.iterator(RDD.scala:287)
    at
org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:96)
    at
org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:53)
    at org.apache.spark.scheduler.Task.run(Task.scala:99)
    at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:282)
    at
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
    at
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
    at java.lang.Thread.run(Thread.java:748)
Caused by: java.lang.ClassNotFoundException:
it.unimi.dsi.fastutil.ints.Int2DoubleOpenHashMap
    at java.net.URLClassLoader.findClass(URLClassLoader.java:382)
    at java.lang.ClassLoader.loadClass(ClassLoader.java:424)
    at java.lang.ClassLoader.loadClass(ClassLoader.java:357)
    ... 35 more
19/03/04 09:13:13 WARN TaskSetManager: Lost task 0.0 in stage 0.0 (TID 1,
localhost, executor driver): java.lang.NoClassDefFoundError:
it/unimi/dsi/fastutil/ints/Int2DoubleOpenHashMap
    at
org.apache.mahout.math.RandomAccessSparseVector.<init>(RandomAccessSparseVector.java:49)
    at
org.apache.mahout.math.RandomAccessSparseVector.<init>(RandomAccessSparseVector.java:44)
    at
org.apache.mahout.sparkbindings.SparkEngine$$anonfun$11$$anonfun$apply$2.apply(SparkEngine.scala:200)
    at
org.apache.mahout.sparkbindings.SparkEngine$$anonfun$11$$anonfun$apply$2.apply(SparkEngine.scala:200)
    at
scala.collection.TraversableLike$$anonfun$map$1.apply(TraversableLike.scala:234)
    at
scala.collection.TraversableLike$$anonfun$map$1.apply(TraversableLike.scala:234)
    at scala.collection.immutable.Range.foreach(Range.scala:160)
    at scala.collection.TraversableLike$class.map(TraversableLike.scala:234)
    at scala.collection.AbstractTraversable.map(Traversable.scala:104)
    at
org.apache.mahout.sparkbindings.SparkEngine$$anonfun$11.apply(SparkEngine.scala:200)
    at
org.apache.mahout.sparkbindings.SparkEngine$$anonfun$11.apply(SparkEngine.scala:195)
    at scala.collection.Iterator$$anon$12.nextCur(Iterator.scala:434)
    at scala.collection.Iterator$$anon$12.hasNext(Iterator.scala:440)
    at scala.collection.Iterator$class.isEmpty(Iterator.scala:330)
    at scala.collection.AbstractIterator.isEmpty(Iterator.scala:1336)
    at
org.apache.mahout.sparkbindings.drm.package$$anonfun$blockify$1.apply(package.scala:55)
    at
org.apache.mahout.sparkbindings.drm.package$$anonfun$blockify$1.apply(package.scala:53)
    at
org.apache.spark.rdd.RDD$$anonfun$mapPartitions$1$$anonfun$apply$23.apply(RDD.scala:796)
    at
org.apache.spark.rdd.RDD$$anonfun$mapPartitions$1$$anonfun$apply$23.apply(RDD.scala:796)
    at
org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:38)
    at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:323)
    at org.apache.spark.rdd.RDD.iterator(RDD.scala:287)
    at
org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:38)
    at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:323)
    at org.apache.spark.rdd.RDD.iterator(RDD.scala:287)
    at
org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:38)
    at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:323)
    at org.apache.spark.rdd.RDD.iterator(RDD.scala:287)
    at
org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:96)
    at
org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:53)
    at org.apache.spark.scheduler.Task.run(Task.scala:99)
    at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:282)
    at
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
    at
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
    at java.lang.Thread.run(Thread.java:748)

19/03/04 09:13:13 ERROR TaskSetManager: Task 0 in stage 0.0 failed 1 times;
aborting job
19/03/04 09:13:13 WARN TaskSetManager: Lost task 0.0 in stage 1.0 (TID 0,
localhost, executor driver): java.lang.NoClassDefFoundError:
it/unimi/dsi/fastutil/ints/Int2DoubleOpenHashMap
    at
org.apache.mahout.math.RandomAccessSparseVector.<init>(RandomAccessSparseVector.java:49)
    at
org.apache.mahout.math.RandomAccessSparseVector.<init>(RandomAccessSparseVector.java:44)
    at
org.apache.mahout.sparkbindings.SparkEngine$$anonfun$11$$anonfun$apply$2.apply(SparkEngine.scala:200)
    at
org.apache.mahout.sparkbindings.SparkEngine$$anonfun$11$$anonfun$apply$2.apply(SparkEngine.scala:200)
    at
scala.collection.TraversableLike$$anonfun$map$1.apply(TraversableLike.scala:234)
    at
scala.collection.TraversableLike$$anonfun$map$1.apply(TraversableLike.scala:234)
    at scala.collection.immutable.Range.foreach(Range.scala:160)
    at scala.collection.TraversableLike$class.map(TraversableLike.scala:234)
    at scala.collection.AbstractTraversable.map(Traversable.scala:104)
    at
org.apache.mahout.sparkbindings.SparkEngine$$anonfun$11.apply(SparkEngine.scala:200)
    at
org.apache.mahout.sparkbindings.SparkEngine$$anonfun$11.apply(SparkEngine.scala:195)
    at scala.collection.Iterator$$anon$12.nextCur(Iterator.scala:434)
    at scala.collection.Iterator$$anon$12.hasNext(Iterator.scala:440)
    at scala.collection.Iterator$class.isEmpty(Iterator.scala:330)
    at scala.collection.AbstractIterator.isEmpty(Iterator.scala:1336)
    at
org.apache.mahout.sparkbindings.drm.package$$anonfun$blockify$1.apply(package.scala:55)
    at
org.apache.mahout.sparkbindings.drm.package$$anonfun$blockify$1.apply(package.scala:53)
    at
org.apache.spark.rdd.RDD$$anonfun$mapPartitions$1$$anonfun$apply$23.apply(RDD.scala:796)
    at
org.apache.spark.rdd.RDD$$anonfun$mapPartitions$1$$anonfun$apply$23.apply(RDD.scala:796)
    at
org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:38)
    at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:323)
    at org.apache.spark.rdd.RDD.iterator(RDD.scala:287)
    at
org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:38)
    at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:323)
    at org.apache.spark.rdd.RDD.iterator(RDD.scala:287)
    at
org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:38)
    at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:323)
    at org.apache.spark.rdd.RDD.iterator(RDD.scala:287)
    at
org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:96)
    at
org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:53)
    at org.apache.spark.scheduler.Task.run(Task.scala:99)
    at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:282)
    at
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
    at
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
    at java.lang.Thread.run(Thread.java:748)
Caused by: java.lang.ClassNotFoundException:
it.unimi.dsi.fastutil.ints.Int2DoubleOpenHashMap
    at java.net.URLClassLoader.findClass(URLClassLoader.java:382)
    at java.lang.ClassLoader.loadClass(ClassLoader.java:424)
    at java.lang.ClassLoader.loadClass(ClassLoader.java:357)
    ... 35 more

19/03/04 09:13:13 ERROR TaskSetManager: Task 0 in stage 1.0 failed 1 times;
aborting job
org.apache.spark.SparkException: Job aborted due to stage failure: Task 0
in stage 0.0 failed 1 times, most recent failure: Lost task 0.0 in stage
0.0 (TID 1, localhost, executor driver): java.lang.NoClassDefFoundError:
it/unimi/dsi/fastutil/ints/Int2DoubleOpenHashMap
    at
org.apache.mahout.math.RandomAccessSparseVector.<init>(RandomAccessSparseVector.java:49)
    at
org.apache.mahout.math.RandomAccessSparseVector.<init>(RandomAccessSparseVector.java:44)
    at
org.apache.mahout.sparkbindings.SparkEngine$$anonfun$11$$anonfun$apply$2.apply(SparkEngine.scala:200)
    at
org.apache.mahout.sparkbindings.SparkEngine$$anonfun$11$$anonfun$apply$2.apply(SparkEngine.scala:200)
    at
scala.collection.TraversableLike$$anonfun$map$1.apply(TraversableLike.scala:234)
    at
scala.collection.TraversableLike$$anonfun$map$1.apply(TraversableLike.scala:234)
    at scala.collection.immutable.Range.foreach(Range.scala:160)
    at scala.collection.TraversableLike$class.map(TraversableLike.scala:234)
    at scala.collection.AbstractTraversable.map(Traversable.scala:104)
    at
org.apache.mahout.sparkbindings.SparkEngine$$anonfun$11.apply(SparkEngine.scala:200)
    at
org.apache.mahout.sparkbindings.SparkEngine$$anonfun$11.apply(SparkEngine.scala:195)
    at scala.collection.Iterator$$anon$12.nextCur(Iterator.scala:434)
    at scala.collection.Iterator$$anon$12.hasNext(Iterator.scala:440)
    at scala.collection.Iterator$class.isEmpty(Iterator.scala:330)
    at scala.collection.AbstractIterator.isEmpty(Iterator.scala:1336)
    at
org.apache.mahout.sparkbindings.drm.package$$anonfun$blockify$1.apply(package.scala:55)
    at
org.apache.mahout.sparkbindings.drm.package$$anonfun$blockify$1.apply(package.scala:53)
    at
org.apache.spark.rdd.RDD$$anonfun$mapPartitions$1$$anonfun$apply$23.apply(RDD.scala:796)
    at
org.apache.spark.rdd.RDD$$anonfun$mapPartitions$1$$anonfun$apply$23.apply(RDD.scala:796)
    at
org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:38)
    at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:323)
    at org.apache.spark.rdd.RDD.iterator(RDD.scala:287)
    at
org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:38)
    at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:323)
    at org.apache.spark.rdd.RDD.iterator(RDD.scala:287)
    at
org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:38)
    at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:323)
    at org.apache.spark.rdd.RDD.iterator(RDD.scala:287)
    at
org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:96)
    at
org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:53)
    at org.apache.spark.scheduler.Task.run(Task.scala:99)
    at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:282)
    at
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
    at
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
    at java.lang.Thread.run(Thread.java:748)

Driver stacktrace:
  at org.apache.spark.scheduler.DAGScheduler.org
$apache$spark$scheduler$DAGScheduler$$failJobAndIndependentStages(DAGScheduler.scala:1435)
  at
org.apache.spark.scheduler.DAGScheduler$$anonfun$abortStage$1.apply(DAGScheduler.scala:1423)
  at
org.apache.spark.scheduler.DAGScheduler$$anonfun$abortStage$1.apply(DAGScheduler.scala:1422)
  at
scala.collection.mutable.ResizableArray$class.foreach(ResizableArray.scala:59)
  at scala.collection.mutable.ArrayBuffer.foreach(ArrayBuffer.scala:48)
  at
org.apache.spark.scheduler.DAGScheduler.abortStage(DAGScheduler.scala:1422)
  at
org.apache.spark.scheduler.DAGScheduler$$anonfun$handleTaskSetFailed$1.apply(DAGScheduler.scala:802)
  at
org.apache.spark.scheduler.DAGScheduler$$anonfun$handleTaskSetFailed$1.apply(DAGScheduler.scala:802)
  at scala.Option.foreach(Option.scala:257)
  at
org.apache.spark.scheduler.DAGScheduler.handleTaskSetFailed(DAGScheduler.scala:802)
  at
org.apache.spark.scheduler.DAGSchedulerEventProcessLoop.doOnReceive(DAGScheduler.scala:1650)
  at
org.apache.spark.scheduler.DAGSchedulerEventProcessLoop.onReceive(DAGScheduler.scala:1605)
  at
org.apache.spark.scheduler.DAGSchedulerEventProcessLoop.onReceive(DAGScheduler.scala:1594)
  at org.apache.spark.util.EventLoop$$anon$1.run(EventLoop.scala:48)
  at org.apache.spark.scheduler.DAGScheduler.runJob(DAGScheduler.scala:628)
  at org.apache.spark.SparkContext.runJob(SparkContext.scala:1918)
  at org.apache.spark.SparkContext.runJob(SparkContext.scala:1931)
  at org.apache.spark.SparkContext.runJob(SparkContext.scala:1944)
  at org.apache.spark.SparkContext.runJob(SparkContext.scala:1958)
  at org.apache.spark.rdd.RDD$$anonfun$collect$1.apply(RDD.scala:935)
  at
org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:151)
  at
org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:112)
  at org.apache.spark.rdd.RDD.withScope(RDD.scala:362)
  at org.apache.spark.rdd.RDD.collect(RDD.scala:934)
  at
org.apache.mahout.sparkbindings.drm.CheckpointedDrmSpark.collect(CheckpointedDrmSpark.scala:128)
  at org.apache.mahout.math.drm.package$.drm2InCore(package.scala:98)
  at timeSparseDRMMMul(<console>:87)
  ... 60 elided
Caused by: java.lang.NoClassDefFoundError:
it/unimi/dsi/fastutil/ints/Int2DoubleOpenHashMap
  at
org.apache.mahout.math.RandomAccessSparseVector.<init>(RandomAccessSparseVector.java:49)
  at
org.apache.mahout.math.RandomAccessSparseVector.<init>(RandomAccessSparseVector.java:44)
  at
org.apache.mahout.sparkbindings.SparkEngine$$anonfun$11$$anonfun$apply$2.apply(SparkEngine.scala:200)
  at
org.apache.mahout.sparkbindings.SparkEngine$$anonfun$11$$anonfun$apply$2.apply(SparkEngine.scala:200)
  at
scala.collection.TraversableLike$$anonfun$map$1.apply(TraversableLike.scala:234)
  at
scala.collection.TraversableLike$$anonfun$map$1.apply(TraversableLike.scala:234)
  at scala.collection.immutable.Range.foreach(Range.scala:160)
  at scala.collection.TraversableLike$class.map(TraversableLike.scala:234)
  at scala.collection.AbstractTraversable.map(Traversable.scala:104)
  at
org.apache.mahout.sparkbindings.SparkEngine$$anonfun$11.apply(SparkEngine.scala:200)
  at
org.apache.mahout.sparkbindings.SparkEngine$$anonfun$11.apply(SparkEngine.scala:195)
  at scala.collection.Iterator$$anon$12.nextCur(Iterator.scala:434)
  at scala.collection.Iterator$$anon$12.hasNext(Iterator.scala:440)
  at scala.collection.Iterator$class.isEmpty(Iterator.scala:330)
  at scala.collection.AbstractIterator.isEmpty(Iterator.scala:1336)
  at
org.apache.mahout.sparkbindings.drm.package$$anonfun$blockify$1.apply(package.scala:55)
  at
org.apache.mahout.sparkbindings.drm.package$$anonfun$blockify$1.apply(package.scala:53)
  at
org.apache.spark.rdd.RDD$$anonfun$mapPartitions$1$$anonfun$apply$23.apply(RDD.scala:796)
  at
org.apache.spark.rdd.RDD$$anonfun$mapPartitions$1$$anonfun$apply$23.apply(RDD.scala:796)
  at
org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:38)
  at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:323)
  at org.apache.spark.rdd.RDD.iterator(RDD.scala:287)
  at
org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:38)
  at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:323)
  at org.apache.spark.rdd.RDD.iterator(RDD.scala:287)
  at
org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:38)
  at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:323)
  at org.apache.spark.rdd.RDD.iterator(RDD.scala:287)
  at
org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:96)
  at
org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:53)
  at org.apache.spark.scheduler.Task.run(Task.scala:99)
  at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:282)
  at
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
  at
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
  at java.lang.Thread.run(Thread.java:748)

scala>


On Mon, Mar 4, 2019 at 7:27 AM Trevor Grant <trevor.d.gr...@gmail.com>
wrote:

> +1 binding.
>
> Steps
> ```
> rm -rf ~/.m2/repository/org/apache/mahout
> wget
>
> https://repository.apache.org/content/repositories/orgapachemahout-1052/org/apache/mahout/mahout/0.14.0/mahout-0.14.0-source-release.zip
> unzip mahout-0.14.0-source-release.zip
> cd mahout-0.14.0
> mvn clean install
> ```
>
> yields-
> ```
> [INFO]
> ------------------------------------------------------------------------
> [INFO] Reactor Summary:
> [INFO]
> [INFO] Apache Mahout ...................................... SUCCESS [
> 2.033 s]
> [INFO] Mahout Core ........................................ SUCCESS [04:06
> min]
> [INFO] Mahout Engine ...................................... SUCCESS [
> 0.122 s]
> [INFO] - Mahout HDFS Support .............................. SUCCESS [
> 10.330 s]
> [INFO] - Mahout Spark Engine .............................. SUCCESS [01:36
> min]
> [INFO] Mahout Community ................................... SUCCESS [
> 0.469 s]
> [INFO]
> ------------------------------------------------------------------------
> [INFO] BUILD SUCCESS
> [INFO]
> ------------------------------------------------------------------------
> [INFO] Total time: 05:56 min
> [INFO] Finished at: 2019-03-04T09:13:09-06:00
> [INFO] Final Memory: 77M/1085M
> [INFO]
> ------------------------------------------------------------------------
> ```
>
> Other info:
> $ java -version
> openjdk version "1.8.0_171"
> OpenJDK Runtime Environment (build
> 1.8.0_171-8u171-b11-0ubuntu0.17.10.1-b11)
> OpenJDK 64-Bit Server VM (build 25.171-b11, mixed mode)
>
> $ mvn -v
> Apache Maven 3.5.0
>
>
>
> On Fri, Mar 1, 2019 at 3:10 PM Andrew Musselman <a...@apache.org> wrote:
>
> > Just remembered md5 is deprecated per ASF release guidelines; sha1 sums
> are
> > good too:
> >
> > bob $ sha1sum mahout-0.14.0.pom
> > dfd3e920e652302823279e01a0a5ab4c819cd54c  mahout-0.14.0.pom
> > bob $ cat mahout-0.14.0.pom.sha1
> > dfd3e920e652302823279e01a0a5ab4c819cd54c
> > bob $ sha1sum mahout-0.14.0-source-release.zip
> > 63daeccdfdd6fc4b2014ad4a35c30c54a08b4a2b
> mahout-0.14.0-source-release.zip
> > bob $ cat mahout-0.14.0-source-release.zip.sha1
> > 63daeccdfdd6fc4b2014ad4a35c30c54a08b4a2b
> > bob $ sha1sum mahout-0.14.0-tests.jar
> > 83d3c495430e4082be4df418b846ab32573e154f  mahout-0.14.0-tests.jar
> > bob $ cat mahout-0.14.0-tests.jar.sha1
> > 83d3c495430e4082be4df418b846ab32573e154f
> >
> > On Fri, Mar 1, 2019 at 11:07 AM Andrew Musselman <a...@apache.org> wrote:
> >
> > > New build out at
> > >
> https://repository.apache.org/content/repositories/orgapachemahout-1052
> > >
> > > Builds, checksums are correct, and signatures are good; I'll be testing
> > > over the weekend.
> > >
> > > [INFO] Apache Mahout ...................................... SUCCESS [
> > > 1.217 s]
> > > [INFO] Mahout Core ........................................ SUCCESS
> > [03:10
> > > min]
> > > [INFO] Mahout Engine ...................................... SUCCESS [
> > > 0.080 s]
> > > [INFO] - Mahout HDFS Support .............................. SUCCESS [
> > > 4.711 s]
> > > [INFO] - Mahout Spark Engine .............................. SUCCESS [
> > > 59.623 s]
> > > [INFO] Mahout Community ................................... SUCCESS [
> > > 0.376 s]
> > > [INFO]
> > >
> ------------------------------------------------------------------------
> > > [INFO] BUILD SUCCESS
> > > [INFO]
> > >
> ------------------------------------------------------------------------
> > > [INFO] Total time: 04:16 min
> > > [INFO] Finished at: 2019-03-01T10:59:01-08:00
> > > [INFO] Final Memory: 71M/1281M
> > > [INFO]
> > >
> ------------------------------------------------------------------------
> > >
> > >
> > > bob $ ls
> > > index.html         mahout-0.14.0.pom.asc
> > > mahout-0.14.0-source-release.zip
> > > mahout-0.14.0-source-release.zip.sha1  mahout-0.14.0-tests.jar.md5
> > > mahout-0.14.0      mahout-0.14.0.pom.md5
> > > mahout-0.14.0-source-release.zip.asc
> > > mahout-0.14.0-tests.jar                mahout-0.14.0-tests.jar.sha1
> > > mahout-0.14.0.pom  mahout-0.14.0.pom.sha1
> > > mahout-0.14.0-source-release.zip.md5  mahout-0.14.0-tests.jar.asc
> > >
> > > bob $ gpg mahout-0.14.0.pom.asc
> > > gpg: assuming signed data in `mahout-0.14.0.pom'
> > > gpg: Signature made Fri 01 Mar 2019 09:59:00 AM PST using RSA key ID
> > > 140A5BE9
> > > gpg: Good signature from "Andrew K Musselman (ASF Signing Key) <
> > > a...@apache.org>"
> > > bob $ gpg mahout-0.14.0-source-release.zip.asc
> > > gpg: assuming signed data in `mahout-0.14.0-source-release.zip'
> > > gpg: Signature made Fri 01 Mar 2019 09:59:00 AM PST using RSA key ID
> > > 140A5BE9
> > > gpg: Good signature from "Andrew K Musselman (ASF Signing Key) <
> > > a...@apache.org>"
> > > bob $ gpg mahout-0.14.0-tests.jar.asc
> > > gpg: assuming signed data in `mahout-0.14.0-tests.jar'
> > > gpg: Signature made Fri 01 Mar 2019 09:59:00 AM PST using RSA key ID
> > > 140A5BE9
> > > gpg: Good signature from "Andrew K Musselman (ASF Signing Key) <
> > > a...@apache.org>"
> > >
> > > bob $ md5sum mahout-0.14.0.pom
> > > 5a2c22802d443eb96afb1afb3f38e9c8  mahout-0.14.0.pom
> > > bob $ cat mahout-0.14.0.pom.md5
> > > 5a2c22802d443eb96afb1afb3f38e9c8
> > > bob $ md5sum mahout-0.14.0-source-release.zip
> > > b6eadad5cdd69f0eccae38f2eebefdd0  mahout-0.14.0-source-release.zip
> > > bob $ cat mahout-0.14.0-source-release.zip.md5
> > > b6eadad5cdd69f0eccae38f2eebefdd0
> > > bob $ md5sum mahout-0.14.0-tests.jar
> > > bef852667898c41fd3c95fef71d6325c  mahout-0.14.0-tests.jar
> > > bob $ cat mahout-0.14.0-tests.jar.md5
> > > bef852667898c41fd3c95fef71d6325c
> > >
> >
>

Reply via email to