I could see "Ticket cache: KEYRING:persistent:1004:1004" in your env.

May be KEYRING persistent cache setting is causing trouble, Kerberos
libraries to store the krb cache in a location and the Hadoop libraries
can't seem to access it.

Please refer these links,
https://community.hortonworks.com/questions/818/ipa-kerberos-not-liking-my-kinit-ticket.html
https://community.hortonworks.com/articles/11291/kerberos-cache-in-ipa-redhat-idm-keyring-solved-1.html


Rakesh
Intel

On Wed, Sep 21, 2016 at 2:21 PM, kevin <kiss.kevin...@gmail.com> wrote:

> [hadoop@dmp1 ~]$ hdfs dfs -ls /
> 16/09/20 15:00:44 WARN ipc.Client: Exception encountered while connecting
> to the server : javax.security.sasl.SaslException: GSS initiate failed
> [Caused by GSSException: No valid credentials provided (Mechanism level:
> Failed to find any Kerberos tgt)]
> ls: Failed on local exception: java.io.IOException: 
> javax.security.sasl.SaslException:
> GSS initiate failed [Caused by GSSException: No valid credentials provided
> (Mechanism level: Failed to find any Kerberos tgt)]; Host Details : local
> host is: "dmp1.youedata.com/192.168.249.129"; destination host is: "
> dmp1.youedata.com":9000;
> [hadoop@dmp1 ~]$ klist
> Ticket cache: KEYRING:persistent:1004:1004
> Default principal: had...@example.com
>
> Valid starting       Expires              Service principal
> 09/20/2016 14:57:34  09/21/2016 14:57:31  krbtgt/example....@example.com
> renew until 09/27/2016 14:57:31
> [hadoop@dmp1 ~]$
>
> I have run kinit had...@example.com before .
>
> 2016-09-21 10:14 GMT+08:00 Wei-Chiu Chuang <weic...@cloudera.com>:
>
>> You need to run kinit command to authenticate before running hdfs dfs -ls
>> command.
>>
>> Wei-Chiu Chuang
>>
>> On Sep 20, 2016, at 6:59 PM, kevin <kiss.kevin...@gmail.com> wrote:
>>
>> Thank you Brahma Reddy Battula.
>> It's because of my problerm of the hdfs-site config file and https
>> ca configuration.
>> now I can startup namenode and I can see the datanodes from the web.
>> but When I try hdfs dfs -ls /:
>>
>> *[hadoop@dmp1 hadoop-2.7.3]$ hdfs dfs -ls /*
>> *16/09/20 07:56:48 WARN ipc.Client: Exception encountered while
>> connecting to the server : javax.security.sasl.SaslException: GSS initiate
>> failed [Caused by GSSException: No valid credentials provided (Mechanism
>> level: Failed to find any Kerberos tgt)]*
>> *ls: Failed on local exception: java.io.IOException:
>> javax.security.sasl.SaslException: GSS initiate failed [Caused by
>> GSSException: No valid credentials provided (Mechanism level: Failed to
>> find any Kerberos tgt)]; Host Details : local host is:
>> "dmp1.example.com/192.168.249.129
>> <http://dmp1.example.com/192.168.249.129>"; destination host is: "dmp1.*
>> *example**.com":9000; *
>>
>> current user is hadoop which startup hdfs , and I have add addprinc
>> hadoop with commond :
>> kadmin.local -q "addprinc hadoop"
>>
>>
>> 2016-09-20 17:33 GMT+08:00 Brahma Reddy Battula <
>> brahmareddy.batt...@huawei.com>:
>>
>>> Seems to be property problem.. it should be *principal* ( ā€œlā€ is
>>> missed).
>>>
>>>
>>>
>>> *<property>*
>>>
>>> *  <name>dfs.secondary.namenode.kerberos.principa</name>*
>>>
>>> *  <value>hadoop/_h...@example.com <h...@example.com></value>*
>>>
>>> *</property>*
>>>
>>>
>>>
>>>
>>>
>>> For namenode httpserver start fail, please check rakesh comments..
>>>
>>>
>>>
>>> This is probably due to some missing configuration.
>>>
>>> Could you please re-check the ssl-server.xml, keystore and truststore
>>> properties:
>>>
>>>
>>>
>>> ssl.server.keystore.location
>>>
>>> ssl.server.keystore.keypassword
>>>
>>> ssl.client.truststore.location
>>>
>>> ssl.client.truststore.password
>>>
>>>
>>>
>>>
>>>
>>> --Brahma Reddy Battula
>>>
>>>
>>>
>>> *From:* kevin [mailto:kiss.kevin...@gmail.com]
>>> *Sent:* 20 September 2016 16:53
>>> *To:* Rakesh Radhakrishnan
>>> *Cc:* user.hadoop
>>> *Subject:* Re: hdfs2.7.3 kerberos can not startup
>>>
>>>
>>>
>>> thanks, but my issue is name node could  *Login successful,but second
>>> namenode couldn't. and name node got a HttpServer.start() threw a non Bind
>>> IOException:*
>>>
>>>
>>>
>>> hdfs-site.xml:
>>>
>>>
>>>
>>> *<property>*
>>>
>>> *    <name>dfs.webhdfs.enabled</name>*
>>>
>>> *    <value>true</value>*
>>>
>>> *</property>*
>>>
>>>
>>>
>>> *<property>*
>>>
>>> *  <name>dfs.block.access.token.enable</name>*
>>>
>>> *  <value>true</value>*
>>>
>>> *</property>*
>>>
>>>
>>>
>>> *<!-- NameNode security config -->*
>>>
>>> *<property>*
>>>
>>> *  <name>dfs.namenode.kerberos.pr
>>> <http://dfs.namenode.kerberos.pr>incipal</name>*
>>>
>>> *  <value>hadoop/_h...@example.com <h...@example.com></value>*
>>>
>>> *</property>*
>>>
>>> *<property>*
>>>
>>> *  <name>dfs.namenode.keytab.file</name>*
>>>
>>> *  <value>/etc/hadoop/conf/hdfs.keytab</value>*
>>>
>>> *</property>*
>>>
>>> *<property>*
>>>
>>> *  <name>dfs.https.port</name>*
>>>
>>> *  <value>50470</value>*
>>>
>>> *</property>*
>>>
>>> *<property>*
>>>
>>> *  <name>dfs.namenode.https-address</name>*
>>>
>>> *  <value>dmp1.example.com:50470
>>> <http://dmp1.example.com:50470/></value>*
>>>
>>> *</property>*
>>>
>>> *<property>*
>>>
>>> *  <name>dfs.namenode.kerberos.in
>>> <http://dfs.namenode.kerberos.in>ternal.spnego.principa</name>*
>>>
>>> *  <value>HTTP/_h...@example.com <h...@example.com></value>*
>>>
>>> *</property>*
>>>
>>> *<property>*
>>>
>>> *  <name>dfs.web.authentication.kerberos.keytab</name>*
>>>
>>> *  <value>/etc/hadoop/conf/hdfs.keytab</value>*
>>>
>>> *</property>*
>>>
>>> *<property>*
>>>
>>> *  <name>dfs.http.policy</name>*
>>>
>>> *  <value>HTTPS_ONLY</value>*
>>>
>>> *</property>*
>>>
>>> *<property>*
>>>
>>> *  <name>dfs.https.enable</name>*
>>>
>>> *  <value>true</value>*
>>>
>>> *</property>*
>>>
>>>
>>>
>>>
>>>
>>> *<!-- secondary NameNode security config -->*
>>>
>>> *<property>*
>>>
>>> *  <name>dfs.namenode.secondary.http-address</name>*
>>>
>>> *  <value>dmp1.example.com:50090
>>> <http://dmp1.example.com:50090/></value>*
>>>
>>> *</property>*
>>>
>>> *<property>*
>>>
>>> *  <name>dfs.secondary.namenode.keytab.file</name>*
>>>
>>> *  <value>/etc/hadoop/conf/hdfs.keytab</value>*
>>>
>>> *</property>*
>>>
>>> *<property>*
>>>
>>> *  <name>dfs.secondary.namenode.kerberos.principa</name>*
>>>
>>> *  <value>hadoop/_h...@example.com <h...@example.com></value>*
>>>
>>> *</property>      *
>>>
>>> *<property>*
>>>
>>> *
>>> <name>dfs.secondary.namenode.kerberos.internal.spnego.principal</name>*
>>>
>>> *  <value>HTTP/_h...@example.com <h...@example.com></value>*
>>>
>>> *</property>*
>>>
>>> *<property>*
>>>
>>> *  <name>dfs.namenode.secondary.https-port</name>*
>>>
>>> *  <value>50470</value>*
>>>
>>> *</property>*
>>>
>>>
>>>
>>>
>>>
>>> *<!-- JournalNode security config -->*
>>>
>>>
>>>
>>> *<property>*
>>>
>>> *  <name>dfs.journalnode.keytab.file</name>*
>>>
>>> *  <value>/etc/hadoop/conf/hdfs.keytab</value>*
>>>
>>> *</property>*
>>>
>>> *<property>*
>>>
>>> *  <name>dfs.journalnode.kerberos.principa</name>*
>>>
>>> *  <value>hadoop/_h...@example.com <h...@example.com></value>*
>>>
>>> *</property>      *
>>>
>>> *<property>*
>>>
>>> *  <name>dfs.journalnode.kerberos.internal.spnego.principa</name>*
>>>
>>> *  <value>HTTP/_h...@example.com <h...@example.com></value>*
>>>
>>> *</property>*
>>>
>>> *<property>*
>>>
>>> *  <name>dfs.web.authentication.kerberos.keytab</name>*
>>>
>>> *  <value>/etc/hadoop/conf/hdfs.keytab</value>*
>>>
>>> *</property>*
>>>
>>>
>>>
>>>
>>>
>>> *<!-- DataNode security config -->*
>>>
>>> *<property>*
>>>
>>> *  <name>dfs.datanode.kerberos.pr
>>> <http://dfs.datanode.kerberos.pr>incipal</name>*
>>>
>>> *  <value>hadoop/_h...@example.com <h...@example.com></value>*
>>>
>>> *</property>*
>>>
>>> *<property>*
>>>
>>> *  <name>dfs.datanode.keytab.file</name>*
>>>
>>> *  <value>/etc/hadoop/conf/hdfs.keytab</value>*
>>>
>>> *</property>*
>>>
>>> *<property>*
>>>
>>> *  <name>dfs.datanode.data.dir.pe
>>> <http://dfs.datanode.data.dir.pe>rm</name>*
>>>
>>> *  <value>700</value>*
>>>
>>> *</property>*
>>>
>>>
>>>
>>> *<!-- datanode SASL-->*
>>>
>>> *<property>*
>>>
>>> *  <name>dfs.datanode.address</name>*
>>>
>>> *  <value>0.0.0.0:61004 <http://0.0.0.0:61004/></value>*
>>>
>>> *</property>*
>>>
>>> *<property>*
>>>
>>> *  <name>dfs.datanode.http.address</name>*
>>>
>>> *  <value>0.0.0.0:61006 <http://0.0.0.0:61006/></value>*
>>>
>>> *</property>*
>>>
>>> *<property>*
>>>
>>> *  <name>dfs.datanode.https.address</name>*
>>>
>>> *  <value>0.0.0.0:50470 <http://0.0.0.0:50470/></value>*
>>>
>>> *</property>*
>>>
>>>
>>>
>>> *<property>*
>>>
>>> *  <name>dfs.data.transfer.protection</name>*
>>>
>>> *  <value>integrity</value>*
>>>
>>> *</property>*
>>>
>>>
>>>
>>> *<property>*
>>>
>>> *     <name>dfs.web.authentication.kerberos.principal</name>*
>>>
>>> *     <value>HTTP/_h...@example.com <h...@example.com></value>*
>>>
>>> *</property>*
>>>
>>> *<property>*
>>>
>>> *     <name>dfs.web.authentication.kerberos.keytab</name>*
>>>
>>> *     <value>/etc/hadoop/conf/hdfs.keytab</value>*
>>>
>>> *</property>*
>>>
>>>
>>>
>>> *and [hadoop@dmp1 hadoop-2.7.3]$ klist -ket /etc/hadoop/conf/hdfs.keytab*
>>>
>>>
>>>
>>>
>>>
>>> *Keytab name: FILE:/etc/hadoop/conf/hdfs.keytab*
>>>
>>> *KVNO Timestamp           Principal*
>>>
>>> *---- -------------------
>>> ------------------------------------------------------*
>>>
>>> *   2 09/19/2016 16:00:41 hdfs/dmp1.example....@example.com
>>> <dmp1.example....@example.com> (aes256-cts-hmac-sha1-96) *
>>>
>>> *   2 09/19/2016 16:00:41 hdfs/dmp1.example....@example.com
>>> <dmp1.example....@example.com> (aes128-cts-hmac-sha1-96) *
>>>
>>> *   2 09/19/2016 16:00:41 hdfs/dmp1.example....@example.com
>>> <dmp1.example....@example.com> (des3-cbc-sha1) *
>>>
>>> *   2 09/19/2016 16:00:41 hdfs/dmp1.example....@example.com
>>> <dmp1.example....@example.com> (arcfour-hmac) *
>>>
>>> *   2 09/19/2016 16:00:41 hdfs/dmp2.example....@example.com
>>> <dmp2.example....@example.com> (aes256-cts-hmac-sha1-96) *
>>>
>>> *   2 09/19/2016 16:00:41 hdfs/dmp2.example....@example.com
>>> <dmp2.example....@example.com> (aes128-cts-hmac-sha1-96) *
>>>
>>> *   2 09/19/2016 16:00:41 hdfs/dmp2.example....@example.com
>>> <dmp2.example....@example.com> (des3-cbc-sha1) *
>>>
>>> *   2 09/19/2016 16:00:41 hdfs/dmp2.example....@example.com
>>> <dmp2.example....@example.com> (arcfour-hmac) *
>>>
>>> *   2 09/19/2016 16:00:41 hdfs/dmp3.example....@example.com
>>> <dmp3.example....@example.com> (aes256-cts-hmac-sha1-96) *
>>>
>>> *   2 09/19/2016 16:00:41 hdfs/dmp3.example....@example.com
>>> <dmp3.example....@example.com> (aes128-cts-hmac-sha1-96) *
>>>
>>> *   2 09/19/2016 16:00:41 hdfs/dmp3.example....@example.com
>>> <dmp3.example....@example.com> (des3-cbc-sha1) *
>>>
>>> *   2 09/19/2016 16:00:41 hdfs/dmp3.example....@example.com
>>> <dmp3.example....@example.com> (arcfour-hmac) *
>>>
>>> *   2 09/19/2016 16:00:41 HTTP/dmp1.example....@example.com
>>> <dmp1.example....@example.com> (aes256-cts-hmac-sha1-96) *
>>>
>>> *   2 09/19/2016 16:00:41 HTTP/dmp1.example....@example.com
>>> <dmp1.example....@example.com> (aes128-cts-hmac-sha1-96) *
>>>
>>> *   2 09/19/2016 16:00:41 HTTP/dmp1.example....@example.com
>>> <dmp1.example....@example.com> (des3-cbc-sha1) *
>>>
>>> *   2 09/19/2016 16:00:41 HTTP/dmp1.example....@example.com
>>> <dmp1.example....@example.com> (arcfour-hmac) *
>>>
>>> *   2 09/19/2016 16:00:41 HTTP/dmp2.example....@example.com
>>> <dmp2.example....@example.com> (aes256-cts-hmac-sha1-96) *
>>>
>>> *   2 09/19/2016 16:00:41 HTTP/dmp2.example....@example.com
>>> <dmp2.example....@example.com> (aes128-cts-hmac-sha1-96) *
>>>
>>> *   2 09/19/2016 16:00:41 HTTP/dmp2.example....@example.com
>>> <dmp2.example....@example.com> (des3-cbc-sha1) *
>>>
>>> *   2 09/19/2016 16:00:41 HTTP/dmp2.example....@example.com
>>> <dmp2.example....@example.com> (arcfour-hmac) *
>>>
>>> *   2 09/19/2016 16:00:41 HTTP/dmp3.example....@example.com
>>> <dmp3.example....@example.com> (aes256-cts-hmac-sha1-96) *
>>>
>>> *   2 09/19/2016 16:00:41 HTTP/dmp3.example....@example.com
>>> <dmp3.example....@example.com> (aes128-cts-hmac-sha1-96) *
>>>
>>> *   2 09/19/2016 16:00:41 HTTP/dmp3.example....@example.com
>>> <dmp3.example....@example.com> (des3-cbc-sha1) *
>>>
>>> *   2 09/19/2016 16:00:41 HTTP/dmp3.example....@example.com
>>> <dmp3.example....@example.com> (arcfour-hmac) *
>>>
>>> *   2 09/19/2016 20:21:03 hadoop/dmp1.example....@example.com
>>> <dmp1.example....@example.com> (aes256-cts-hmac-sha1-96) *
>>>
>>> *   2 09/19/2016 20:21:03 hadoop/dmp1.example....@example.com
>>> <dmp1.example....@example.com> (aes128-cts-hmac-sha1-96) *
>>>
>>> *   2 09/19/2016 20:21:03 hadoop/dmp1.example....@example.com
>>> <dmp1.example....@example.com> (des3-cbc-sha1) *
>>>
>>> *   2 09/19/2016 20:21:03 hadoop/dmp1.example....@example.com
>>> <dmp1.example....@example.com> (arcfour-hmac) *
>>>
>>> *   2 09/19/2016 20:21:03 hadoop/dmp2.example....@example.com
>>> <dmp2.example....@example.com> (aes256-cts-hmac-sha1-96) *
>>>
>>> *   2 09/19/2016 20:21:03 hadoop/dmp2.example....@example.com
>>> <dmp2.example....@example.com> (aes128-cts-hmac-sha1-96) *
>>>
>>> *   2 09/19/2016 20:21:03 hadoop/dmp2.example....@example.com
>>> <dmp2.example....@example.com> (des3-cbc-sha1) *
>>>
>>> *   2 09/19/2016 20:21:03 hadoop/dmp2.example....@example.com
>>> <dmp2.example....@example.com> (arcfour-hmac) *
>>>
>>> *   2 09/19/2016 20:21:03 hadoop/dmp3.example....@example.com
>>> <dmp3.example....@example.com> (aes256-cts-hmac-sha1-96) *
>>>
>>> *   2 09/19/2016 20:21:03 hadoop/dmp3.example....@example.com
>>> <dmp3.example....@example.com> (aes128-cts-hmac-sha1-96) *
>>>
>>> *   2 09/19/2016 20:21:03 hadoop/dmp3.example....@example.com
>>> <dmp3.example....@example.com> (des3-cbc-sha1) *
>>>
>>> *   2 09/19/2016 20:21:03 hadoop/dmp3.example....@example.com
>>> <dmp3.example....@example.com> (arcfour-hmac) *
>>>
>>>
>>>
>>> 2016-09-20 15:52 GMT+08:00 Rakesh Radhakrishnan <rake...@apache.org>:
>>>
>>> >>>>>>Caused by: javax.security.auth.login.LoginException: Unable to
>>> obtain password from user
>>>
>>>
>>>
>>> Could you please check kerberos principal name is specified correctly in
>>>
>>> "hdfs-site.xml", which is used to authenticate against Kerberos.
>>>
>>>
>>>
>>> If keytab file defined in "hdfs-site.xml" and doesn't exists or wrong
>>> path, you will see
>>>
>>> this error. So, please verify the path and the keytab filename correctly
>>>
>>> configured.
>>>
>>>
>>>
>>> I hope hadoop discussion thread, https://goo.gl/M6l3vv may help you.
>>>
>>>
>>>
>>>
>>>
>>> >>>>>>>2016-09-20 00:54:06,665 INFO org.apache.hadoop.http.HttpServer2:
>>> HttpServer.start() threw a non Bind IOException
>>>
>>> java.io.IOException: !JsseListener: java.lang.NullPointerException
>>>
>>>
>>>
>>> This is probably due to some missing configuration.
>>>
>>> Could you please re-check the ssl-server.xml, keystore and truststore
>>> properties:
>>>
>>>
>>>
>>> ssl.server.keystore.location
>>>
>>> ssl.server.keystore.keypassword
>>>
>>> ssl.client.truststore.location
>>>
>>> ssl.client.truststore.password
>>>
>>>
>>>
>>> Rakesh
>>>
>>>
>>>
>>> On Tue, Sep 20, 2016 at 10:53 AM, kevin <kiss.kevin...@gmail.com> wrote:
>>>
>>> *hi,all:*
>>>
>>> *My environment : Centos7.2 hadoop2.7.3 jdk1.8*
>>>
>>> *after I config hdfs with kerberos ,I can't start up with
>>> sbin/start-dfs.sh*
>>>
>>>
>>>
>>> *::namenode log as below  *
>>>
>>>
>>>
>>> *STARTUP_MSG:   build = Unknown -r Unknown; compiled by 'root' on
>>> 2016-09-18T09:05Z*
>>>
>>> *STARTUP_MSG:   java = 1.8.0_102*
>>>
>>> *************************************************************/*
>>>
>>> *2016-09-20 00:54:05,822 INFO
>>> org.apache.hadoop.hdfs.server.namenode.NameNode: registered UNIX signal
>>> handlers for [TERM, HUP, INT]*
>>>
>>> *2016-09-20 00:54:05,825 INFO
>>> org.apache.hadoop.hdfs.server.namenode.NameNode: createNameNode []*
>>>
>>> *2016-09-20 00:54:06,078 INFO
>>> org.apache.hadoop.metrics2.impl.MetricsConfig: loaded properties from
>>> hadoop-metrics2.properties*
>>>
>>> *2016-09-20 00:54:06,149 INFO
>>> org.apache.hadoop.metrics2.impl.MetricsSystemImpl: Scheduled snapshot
>>> period at 10 second(s).*
>>>
>>> *2016-09-20 00:54:06,149 INFO
>>> org.apache.hadoop.metrics2.impl.MetricsSystemImpl: NameNode metrics system
>>> started*
>>>
>>> *2016-09-20 00:54:06,151 INFO
>>> org.apache.hadoop.hdfs.server.namenode.NameNode: fs.defaultFS is
>>> hdfs://dmp1.example.com:9000 <http://dmp1.example.com:9000/>*
>>>
>>> *2016-09-20 00:54:06,152 INFO
>>> org.apache.hadoop.hdfs.server.namenode.NameNode: Clients are to use
>>> dmp1.example.com:9000 <http://dmp1.example.com:9000/> to access this
>>> namenode/service.*
>>>
>>> *2016-09-20 00:54:06,446 INFO
>>> org.apache.hadoop.security.UserGroupInformation: Login successful for user
>>> hadoop/dmp1.example....@example.com <dmp1.example....@example.com> using
>>> keytab file /etc/hadoop/conf/hdfs.keytab*
>>>
>>> *2016-09-20 00:54:06,472 INFO org.apache.hadoop.hdfs.DFSUtil: Starting
>>> web server as: HTTP/dmp1.example....@example.com
>>> <dmp1.example....@example.com>*
>>>
>>> *2016-09-20 00:54:06,475 INFO org.apache.hadoop.hdfs.DFSUtil: Starting
>>> Web-server for hdfs at: https://dmp1.example.com:50470
>>> <https://dmp1.example.com:50470/>*
>>>
>>> *2016-09-20 00:54:06,517 INFO org.mortbay.log: Logging to
>>> org.slf4j.impl.Log4jLoggerAdapter(org.mortbay.log) via
>>> org.mortbay.log.Slf4jLog*
>>>
>>> *2016-09-20 00:54:06,533 INFO
>>> org.apache.hadoop.security.authentication.server.AuthenticationFilter:
>>> Unable to initialize FileSignerSecretProvider, falling back to use random
>>> secrets.*
>>>
>>> *2016-09-20 00:54:06,542 INFO org.apache.hadoop.http.HttpRequestLog:
>>> Http request log for http.requests.namenode is not defined*
>>>
>>> *2016-09-20 00:54:06,546 INFO org.apache.hadoop.http.HttpServer2: Added
>>> global filter 'safety'
>>> (class=org.apache.hadoop.http.HttpServer2$QuotingInputFilter)*
>>>
>>> *2016-09-20 00:54:06,548 INFO org.apache.hadoop.http.HttpServer2: Added
>>> filter static_user_filter
>>> (class=org.apache.hadoop.http.lib.StaticUserWebFilter$StaticUserFilter) to
>>> context hdfs*
>>>
>>> *2016-09-20 00:54:06,548 INFO org.apache.hadoop.http.HttpServer2: Added
>>> filter static_user_filter
>>> (class=org.apache.hadoop.http.lib.StaticUserWebFilter$StaticUserFilter) to
>>> context static*
>>>
>>> *2016-09-20 00:54:06,548 INFO org.apache.hadoop.http.HttpServer2: Added
>>> filter static_user_filter
>>> (class=org.apache.hadoop.http.lib.StaticUserWebFilter$StaticUserFilter) to
>>> context logs*
>>>
>>> *2016-09-20 00:54:06,653 INFO org.apache.hadoop.http.HttpServer2: Added
>>> filter 'org.apache.hadoop.hdfs.web.Au
>>> <http://org.apache.hadoop.hdfs.web.au/>thFilter'
>>> (class=org.apache.hadoop.hdfs.web.AuthFilter)*
>>>
>>> *2016-09-20 00:54:06,654 INFO org.apache.hadoop.http.HttpServer2:
>>> addJerseyResourcePackage:
>>> packageName=org.apache.hadoop.hdfs.server.namenode.web.resources;org.apache.hadoop.hdfs.web.resources,
>>> pathSpec=/webhdfs/v1/**
>>>
>>> *2016-09-20 00:54:06,657 INFO org.apache.hadoop.http.HttpServer2: Adding
>>> Kerberos (SPNEGO) filter to getDelegationToken*
>>>
>>> *2016-09-20 00:54:06,658 INFO org.apache.hadoop.http.HttpServer2: Adding
>>> Kerberos (SPNEGO) filter to renewDelegationToken*
>>>
>>> *2016-09-20 00:54:06,658 INFO org.apache.hadoop.http.HttpServer2: Adding
>>> Kerberos (SPNEGO) filter to cancelDelegationToken*
>>>
>>> *2016-09-20 00:54:06,659 INFO org.apache.hadoop.http.HttpServer2: Adding
>>> Kerberos (SPNEGO) filter to fsck*
>>>
>>> *2016-09-20 00:54:06,659 INFO org.apache.hadoop.http.HttpServer2: Adding
>>> Kerberos (SPNEGO) filter to imagetransfer*
>>>
>>> *2016-09-20 00:54:06,665 WARN org.mortbay.log:
>>> java.lang.NullPointerException*
>>>
>>> *2016-09-20 00:54:06,665 INFO org.apache.hadoop.http.HttpServer2:
>>> HttpServer.start() threw a non Bind IOException*
>>>
>>> *java.io.IOException: !JsseListener: java.lang.NullPointerException*
>>>
>>> *at
>>> org.mortbay.jetty.security.SslSocketConnector.newServerSocket(SslSocketConnector.java:516)*
>>>
>>> *at
>>> org.apache.hadoop.security.ssl.SslSocketConnectorSecure.newServerSocket(SslSocketConnectorSecure.java:47)*
>>>
>>> *at org.mortbay.jetty.bio.SocketConnector.open(SocketConnector.java:73)*
>>>
>>> *at
>>> org.apache.hadoop.http.HttpServer2.openListeners(HttpServer2.java:914)*
>>>
>>> *at org.apache.hadoop.http.HttpServer2.start(HttpServer2.java:856)*
>>>
>>> *at org.apache.hadoop.hdfs.server.namenode.NameNodeHttpServer.st
>>> <http://namenode.namenodehttpserver.st/>art(NameNodeHttpServer.java:142)*
>>>
>>> *at
>>> org.apache.hadoop.hdfs.server.namenode.NameNode.startHttpServer(NameNode.java:753)*
>>>
>>> *at
>>> org.apache.hadoop.hdfs.server.namenode.NameNode.initialize(NameNode.java:639)*
>>>
>>> *at
>>> org.apache.hadoop.hdfs.server.namenode.NameNode.<init>(NameNode.java:812)*
>>>
>>> *at
>>> org.apache.hadoop.hdfs.server.namenode.NameNode.<init>(NameNode.java:796)*
>>>
>>> *at
>>> org.apache.hadoop.hdfs.server.namenode.NameNode.createNameNode(NameNode.java:1493)*
>>>
>>> *at
>>> org.apache.hadoop.hdfs.server.namenode.NameNode.main(NameNode.java:1559)*
>>>
>>>
>>>
>>>
>>>
>>> *::second namenode log as below  *
>>>
>>>
>>>
>>> *STARTUP_MSG:   build = Unknown -r Unknown; compiled by 'root' on
>>> 2016-09-18T09:05Z*
>>>
>>> *STARTUP_MSG:   java = 1.8.0_102*
>>>
>>> *************************************************************/*
>>>
>>> *2016-09-20 00:54:14,885 INFO
>>> org.apache.hadoop.hdfs.server.namenode.SecondaryNameNode: registered UNIX
>>> signal handlers for [TERM, HUP, INT]*
>>>
>>> *2016-09-20 00:54:15,263 FATAL
>>> org.apache.hadoop.hdfs.server.namenode.SecondaryNameNode: Failed to start
>>> secondary namenode*
>>>
>>> *java.io.IOException: Login failure for hadoop from keytab
>>> /etc/hadoop/conf/hdfs.keytab: javax.security.auth.login.LoginException:
>>> Unable to obtain password from user*
>>>
>>>
>>>
>>> *at
>>> org.apache.hadoop.security.UserGroupInformation.loginUserFromKeytab(UserGroupInformation.java:963)*
>>>
>>> *at org.apache.hadoop.security.SecurityUtil.login(SecurityUtil.java:246)*
>>>
>>> *at
>>> org.apache.hadoop.hdfs.server.namenode.SecondaryNameNode.initialize(SecondaryNameNode.java:217)*
>>>
>>> *at
>>> org.apache.hadoop.hdfs.server.namenode.SecondaryNameNode.<init>(SecondaryNameNode.java:192)*
>>>
>>> *at
>>> org.apache.hadoop.hdfs.server.namenode.SecondaryNameNode.main(SecondaryNameNode.java:671)*
>>>
>>> *Caused by: javax.security.auth.login.LoginException: Unable to obtain
>>> password from user*
>>>
>>>
>>>
>>> *at
>>> com.sun.security.auth.module.Krb5LoginModule.promptForPass(Krb5LoginModule.java:897)*
>>>
>>> *at
>>> com.sun.security.auth.module.Krb5LoginModule.attemptAuthentication(Krb5LoginModule.java:760)*
>>>
>>> *at
>>> com.sun.security.auth.module.Krb5LoginModule.login(Krb5LoginModule.java:617)*
>>>
>>> *at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)*
>>>
>>> *at
>>> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)*
>>>
>>> *at
>>> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)*
>>>
>>> *at java.lang.reflect.Method.invoke(Method.java:498)*
>>>
>>> *at javax.security.auth.login.LoginContext.invoke(LoginContext.java:755)*
>>>
>>> *at
>>> javax.security.auth.login.LoginContext.access$000(LoginContext.java:195)*
>>>
>>> *at javax.security.auth.login.LoginContext$4.run(LoginContext.java:682)*
>>>
>>> *at javax.security.auth.login.LoginContext$4.run(LoginContext.java:680)*
>>>
>>> *at java.security.AccessController.doPrivileged(Native Method)*
>>>
>>> *at
>>> javax.security.auth.login.LoginContext.invokePriv(LoginContext.java:680)*
>>>
>>> *at javax.security.auth.login.LoginContext.login(LoginContext.java:587)*
>>>
>>> *at
>>> org.apache.hadoop.security.UserGroupInformation.loginUserFromKeytab(UserGroupInformation.java:954)*
>>>
>>> *... 4 more*
>>>
>>>
>>>
>>>
>>>
>>
>>
>>
>

Reply via email to