[
https://issues.apache.org/jira/browse/HIVE-13651?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15296662#comment-15296662
]
Sergey Shelukhin commented on HIVE-13651:
-----------------------------------------
[~jdere] [~sseth] how are these credentials normally provided to a Tez job? Is
this about HDFS/HBase/etc tokens - I assume HS2 issues those, right?
How does SparkSQL currently get HDFS access to read directly from Hive (or
HBase, for example)
> LlapBaseInputFormat: figure out where credentials come from
> -----------------------------------------------------------
>
> Key: HIVE-13651
> URL: https://issues.apache.org/jira/browse/HIVE-13651
> Project: Hive
> Issue Type: Sub-task
> Components: llap
> Reporter: Jason Dere
>
> todo in LlapBaseInputFormat.constructSubmitWorkRequestProto()
> {code}
> // TODO Figure out where credentials will come from. Normally Hive sets up
> // URLs on the tez dag, for which Tez acquires credentials.
> // taskCredentials.addAll(getContext().getCredentials());
> //
> Preconditions.checkState(currentQueryIdentifierProto.getDagIdentifier() ==
> //
> taskSpec.getTaskAttemptID().getTaskID().getVertexID().getDAGId().getId());
> // ByteBuffer credentialsBinary =
> credentialMap.get(currentQueryIdentifierProto);
> // if (credentialsBinary == null) {
> // credentialsBinary =
> serializeCredentials(getContext().getCredentials());
> // credentialMap.putIfAbsent(currentQueryIdentifierProto,
> credentialsBinary.duplicate());
> // } else {
> // credentialsBinary = credentialsBinary.duplicate();
> // }
> //
> builder.setCredentialsBinary(ByteString.copyFrom(credentialsBinary));
> {code}
--
This message was sent by Atlassian JIRA
(v6.3.4#6332)