I just tried using AccumuloInputFormat as a data source for Spark running in standalone mode on a single node 'cluster'. Everything seems to work fine out of the box, as advertised. (Spark is supposed to work with any hadoop InputFormat)
Just properly configure the AccumuloInputFormat, and pass it off to JavaSparkContext.newAPIHadoopRDD(...) to load the data into an RDD. The versions I tested with were Accumulo 1.5, Hadoop 1.2.1, and Spark 0.8.1. Is anyone else using Spark with Accumulo?
