[
https://issues.apache.org/jira/browse/YARN-4721?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15184752#comment-15184752
]
Steve Loughran commented on YARN-4721:
--------------------------------------
I'm trying to say, the test is "if this cluster has a default filesystem, it
can be listed"
not HDFS, just that fs.default.name is not empty. We could even make the check
conditional on the cluster having a default FS. But if you do have a default
FS, YARN had better be able to talk to it. I will change the title to make
clear that this is broader than just HDFS.
> One could argue for a stand-alone service (outside of YARN) that does these
> validations.
That doesn't address the problem I'm looking at, which is: validate that a
specific process started under a specific principal on a specific host has the
credentials needed to access a critical part of the cluster infrastructure, the
default FS.
> So, the notion of "this cluster cannot talk to my HDFS" doesn't generalize.
> It is context dependent and almost always "may app cannot talk to this and
> that HDFS instances".
I agree, which is why distcp will need special attention. However, YARN does
have a specific notion of the defaultFS for a filesystem; ATS1.5 ramps it by
only working with an FS which implements flush() by making data durable and
visible to others (though it doesn't require the metadata to be
complete/visible).
It's authentication of that YARN process to the cluster FS —or more
specifically, identifying why it sometimes doesn't happen— that I'm trying to
look at.
Anyway, this initial patch doesn't attempt any of that, it looks at
UGI.isSecurityEnabled, and if so does some extra diagnostics, fails fast on a
few conditions guaranteed to stop hadoop working. Do you have any issues with
that part of the patch?
> RM to try to auth with HDFS on startup, retry with max diagnostics on failure
> -----------------------------------------------------------------------------
>
> Key: YARN-4721
> URL: https://issues.apache.org/jira/browse/YARN-4721
> Project: Hadoop YARN
> Issue Type: Improvement
> Components: resourcemanager
> Affects Versions: 2.8.0
> Reporter: Steve Loughran
> Assignee: Steve Loughran
> Attachments: HADOOP-12889-001.patch
>
>
> If the RM can't auth with HDFS, this can first surface during job submission,
> which can cause confusion about what's wrong and whose credentials are
> playing up.
> Instead, the RM could try to talk to HDFS on launch, {{ls /}} should suffice.
> If it can't auth, it can then tell UGI to log more and retry.
> I don't know what the policy should be if the RM can't auth to HDFS at this
> point. Certainly it can't currently accept work. But should it fail fast or
> keep going in the hope that the problem is in the KDC or NN and will fix
> itself without an RM restart?
--
This message was sent by Atlassian JIRA
(v6.3.4#6332)