Has anyone used the python Spark API and AccumuloInputFormat?

Using AccumuloInputFormat in scala and java within spark is
straightforward, but the python spark API's newAPIHadoopRDD function takes
in its configuration via a python dict (
https://spark.apache.org/docs/1.1.0/api/python/pyspark.context.SparkContext-class.html#newAPIHadoopRDD)
and there isn't an obvious job configuration set of keys to use. From
looking at the Accumulo source, it seems job configuration values are
stored with keys that are java enums and it's unclear to me what to use for
configuration keys in my python dict.

Any thoughts as to how to do this would be helpful!

Thanks,

Kina

Reply via email to