[snip]
>> I can't call AccumuloInputFormat.setConnectorInfo again since it has
>> already been called, and I presume adding the serialised token to the
>> Configuration would be insecure?
> Yeah, the configuration can't protect sensitive information. MapReduce/YARN
> has special handling to make sure those tokens serialized in the Job's
> credentials are only readable by you (the job submitter).
>
> The thing I don't entirely follow is how you've gotten into this situation
> to begin with. The adding of the delegation tokens to the Job's credentials
> should be done by Accumulo's MR code on your behalf (just like it's
> obtaining the delegation token, it would automatically add it to the job for
> ya).
>
> Any chance you can provide an end-to-end example? I am also pretty
> Spark-ignorant -- so maybe I just don't understand what is possible and what
> isn't..

Hmm, after further investigation concentrating on just MapReduce (and
not Spark) it seems the GeoMesaAccumuloInputFormat class might need
more significant work than just s/PasswordToken/KerberosToken that I
got away with previously. For example, sending an Accumulo password in
the Hadoop conf probably isn't ideal either.

Fortunately I found this:

https://github.com/apache/hive/blob/master/accumulo-handler/src/java/org/apache/hadoop/hive/accumulo/mr/HiveAccumuloTableInputFormat.java

Is it a good example of Accumulo + MapReduce that I can copy?

Thanks,

James

Reply via email to