Thanks Jeff.

But the link above has the HA URI fs.defaultFS
<https://cwiki.apache.org/confluence/display/AMBARI/fs.defaultFS>=hdfs://mycluster
in the HA hdfs blueprint. Any idea why it does not have port? Thanks.

On Wed, Jul 18, 2018 at 2:07 PM, Jeff Hubbs <jhubbsl...@att.net> wrote:

> Lian -
>
> Your value of fs.defaultFS is supposed to have a port number, e.g.
> hdfs://test-cluster:9000.
>
> On 7/18/18 4:28 PM, Lian Jiang wrote:
>
> Hi,
>
> I am enabling HA for hdfs and yarn on my hdp2.6 cluster. HDFS can start
> but yarn cannot due to error:
>
> 2018-07-18 18:11:23,967 FATAL 
> applicationhistoryservice.ApplicationHistoryServer
> (ApplicationHistoryServer.java:launchAppHistoryServer(177)) - Error
> starting ApplicationHistoryServer
> java.lang.IllegalArgumentException: java.net.UnknownHostException:
> test-cluster
>         at org.apache.hadoop.security.SecurityUtil.buildTokenService(Se
> curityUtil.java:439)
>         at org.apache.hadoop.hdfs.NameNodeProxies.createNonHAProxy(Name
> NodeProxies.java:321)
>         at org.apache.hadoop.hdfs.NameNodeProxies.createProxy(NameNodeP
> roxies.java:176)
>         at org.apache.hadoop.hdfs.DFSClient.<init>(DFSClient.java:690)
>         at org.apache.hadoop.hdfs.DFSClient.<init>(DFSClient.java:631)
>         at org.apache.hadoop.hdfs.DistributedFileSystem.initialize(Dist
> ributedFileSystem.java:160)
>         at org.apache.hadoop.fs.FileSystem.createFileSystem(FileSystem.
> java:2795)
>         at org.apache.hadoop.fs.FileSystem.access$200(FileSystem.java:99)
>         at org.apache.hadoop.fs.FileSystem$Cache.getInternal(FileSystem
> .java:2829)
>         at org.apache.hadoop.fs.FileSystem$Cache.get(FileSystem.java:2811)
>         at org.apache.hadoop.fs.FileSystem.get(FileSystem.java:390)
>         at org.apache.hadoop.fs.FileSystem.get(FileSystem.java:179)
>         at org.apache.hadoop.fs.FileSystem.get(FileSystem.java:374)
>         at org.apache.hadoop.fs.Path.getFileSystem(Path.java:295)
>         at org.apache.hadoop.yarn.server.timeline.EntityGroupFSTimeline
> Store.serviceInit(EntityGroupFSTimelineStore.java:209)
>         at org.apache.hadoop.service.AbstractService.init(AbstractServi
> ce.java:163)
>         at org.apache.hadoop.service.CompositeService.serviceInit(Compo
> siteService.java:107)
>         at org.apache.hadoop.yarn.server.applicationhistoryservice.Appl
> icationHistoryServer.serviceInit(ApplicationHistoryServer.java:111)
>         at org.apache.hadoop.service.AbstractService.init(AbstractServi
> ce.java:163)
>         at org.apache.hadoop.yarn.server.applicationhistoryservice.Appl
> icationHistoryServer.launchAppHistoryServer(ApplicationHisto
> ryServer.java:174)
>         at org.apache.hadoop.yarn.server.applicationhistoryservice.Appl
> icationHistoryServer.main(ApplicationHistoryServer.java:184)
>
>
> My current fs.defaultFS is "hdfs://test-cluster" - the HA hdfs uri. Looks
> like yarn cannot recognize this uri.
>
> I found an example at https://cwiki.apache.org/confl
> uence/display/AMBARI/Blueprint+Support+for+HA+Clusters which uses
> fs.defaultFS=hdfs://%HOSTGROUP::master_1%:8020 which is not the HA hdfs
> uri. Plus, this example uses different blueprints for HA hdfs and HA yarn
> while I use one blueprint for both.
>
> My questions:
> 1. should I separate the blueprints for HA Hdfs and HA yarn as above
> example? Or one blueprint for both that I am currently using is fine?
>
> 2. can yarn use the HA hdfs uri instead of pointing to a specific
> namenode? The latter seems to break HA.
>
> Appreciate your hints.
>
>
>

Reply via email to