Following slider specific properties are at present added in
/data/slider/conf/slider-client.xml. If you think they should be picked up
from HADOOP_CONF_DIR (/etc/hadoop/conf) file, which file in HADOOP_CONF_DIR
should these be added ?

   - slider.zookeeper.quorum
   - hadoop.registry.zk.quorum
   - hadoop.registry.zk.root
   - slider.yarn.queue


On Thu, Jul 28, 2016 at 4:37 PM, Gour Saha <gs...@hortonworks.com> wrote:

> That is strange, since it is indeed not required to contain anything in
> slider-client.xml (except <configuration></configuration>) if
> HADOOP_CONF_DIR has everything that Slider needs. This probably gives an
> indication that there might be some issue with cluster configuration based
> on files solely under HADOOP_CONF_DIR to begin with.
>
> Suggest you to upload all the config files to the jira to help debug this
> further.
>
> -Gour
>
> On 7/28/16, 4:27 PM, "Manoj Samel" <manojsamelt...@gmail.com> wrote:
>
> >Thanks Gour for prompt reply
> >
> >BTW - Creating a empty slider-client.xml (with just
> ><configuration></configuration>) does not works. The AM starts but fails
> >to
> >create any components and shows errors like
> >
> >2016-07-28 23:18:46,018
> >[AmExecutor-006-SendThread(localhost.localdomain:2181)] WARN
> > zookeeper.ClientCnxn - Session 0x0 for server null, unexpected error,
> >closing socket connection and attempting reconnect
> >java.net.ConnectException: Connection refused
> >        at sun.nio.ch.SocketChannelImpl.checkConnect(Native Method)
> >        at
> >sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:717)
> >        at
> >org.apache.zookeeper.ClientCnxnSocketNIO.doTransport(ClientCnxnSocketNIO.j
> >ava:361)
> >        at
> >org.apache.zookeeper.ClientCnxn$SendThread.run(ClientCnxn.java:1081)
> >
> >Also, command "slider destroy <app>" fails with zookeeper errors ...
> >
> >I had to keep a minimal slider-client.xml. It does not have any RM info
> >etc. but does contain slider ZK related properties like
> >"slider.zookeeper.quorum", "hadoop.registry.zk.quorum",
> >"hadoop.registry.zk.root". I haven't yet distilled the absolute minimal
> >set
> >of properties required, but this should suffice for now. All RM / HDFS
> >properties will be read from HADOOP_CONF_DIR files.
> >
> >Let me know if this could cause any issues.
> >
> >On Thu, Jul 28, 2016 at 3:36 PM, Gour Saha <gs...@hortonworks.com> wrote:
> >
> >> No need to copy any files. Pointing HADOOP_CONF_DIR to /etc/hadoop/conf
> >>is
> >> good.
> >>
> >> -Gour
> >>
> >> On 7/28/16, 3:24 PM, "Manoj Samel" <manojsamelt...@gmail.com> wrote:
> >>
> >> >Follow up question regarding Gour's comment in earlier thread -
> >> >
> >> >Slider is installed on one of the hadoop nodes. SLIDER_HOME/conf
> >>directory
> >> >(say /data/slider/conf) is different than HADOOP_CONF_DIR
> >> >(/etc/hadoop/conf). Is it required/recommended that files in
> >> >HADOOP_CONF_DIR be copied to SLIDER_HOME/conf and slider-env.sh script
> >> >sets
> >> >HADOOP_CONF_DIR to /data/slider/conf ?
> >> >
> >> >Or can the slider-env.sh set HADOOP_CONF_DIR to /etc/hadoop/conf ,
> >>without
> >> >copying the files ?
> >> >
> >> >Using slider .80 for now, but would like to know recommendation for
> >>this
> >> >and future versions as well.
> >> >
> >> >Thanks in advance,
> >> >
> >> >Manoj
> >> >
> >> >On Tue, Jul 26, 2016 at 3:27 PM, Manoj Samel <manojsamelt...@gmail.com
> >
> >> >wrote:
> >> >
> >> >> Filed https://issues.apache.org/jira/browse/SLIDER-1158 with logs
> and
> >> my
> >> >> analysis of logs.
> >> >>
> >> >> On Tue, Jul 26, 2016 at 10:36 AM, Gour Saha <gs...@hortonworks.com>
> >> >>wrote:
> >> >>
> >> >>> Please file a JIRA and upload the logs to it.
> >> >>>
> >> >>> On 7/26/16, 10:21 AM, "Manoj Samel" <manojsamelt...@gmail.com>
> >>wrote:
> >> >>>
> >> >>> >Hi Gour,
> >> >>> >
> >> >>> >Can you please reach me using your own email-id? I will then send
> >> >>>logs to
> >> >>> >you, along with my analysis - I don't want to send logs on public
> >>list
> >> >>> >
> >> >>> >Thanks,
> >> >>> >
> >> >>> >On Mon, Jul 25, 2016 at 5:39 PM, Gour Saha <gs...@hortonworks.com>
> >> >>> wrote:
> >> >>> >
> >> >>> >> Ok, so this node is not a gateway. It is part of the cluster,
> >>which
> >> >>> >>means
> >> >>> >> you don¹t need slider-client.xml at all. Just have
> >>HADOOP_CONF_DIR
> >> >>> >> pointing to /etc/hadoop/conf in slider-env.sh and that should be
> >>it.
> >> >>> >>
> >> >>> >> So the above simplifies your config setup. It will not solve
> >>either
> >> >>>of
> >> >>> >>the
> >> >>> >> 2 problems you are facing.
> >> >>> >>
> >> >>> >> Now coming to the 2 issues you are facing, you have to provide
> >> >>> >>additional
> >> >>> >> logs for us to understand better. Let¹s start with  -
> >> >>> >> 1. RM logs (specifically between the time when rm1->rm2 failover
> >>is
> >> >>> >> simulated)
> >> >>> >> 2. Slider App logs
> >> >>> >>
> >> >>> >> -Gour
> >> >>> >>
> >> >>> >> On 7/25/16, 5:16 PM, "Manoj Samel" <manojsamelt...@gmail.com>
> >> wrote:
> >> >>> >>
> >> >>> >> >   1. Not clear about your question on "gateway" node. The node
> >> >>> running
> >> >>> >> >   slider is part of the hadoop cluster and there are other
> >> >>>services
> >> >>> >>like
> >> >>> >> >   Oozie that run on this node that utilizes hdfs and yarn. So
> >>if
> >> >>>your
> >> >>> >> >   question is whether the node is otherwise working for HDFS
> >>and
> >> >>>Yarn
> >> >>> >> >   configuration, it is working
> >> >>> >> >   2. I copied all files from HADOOP_CONF_DIR (say
> >> >>>/etc/hadoop/conf)
> >> >>> to
> >> >>> >> >the
> >> >>> >> >   directory containing slider-client.xml (say
> >>/data/latest/conf)
> >> >>> >> >   3. In earlier email, I had done a mistake where slider-env.sh
> >> >>>file
> >> >>> >> >HADOOP_CONF_DIR
> >> >>> >> >   was pointing to original directory /etc/hadoop/conf. I edited
> >> >>>it to
> >> >>> >> >   point to same directory containing slider-client.xml &
> >> >>> slider-env.sh
> >> >>> >> >i.e.
> >> >>> >> >   /data/latest/conf
> >> >>> >> >   4. I emptied slider-client.xml. It just had the
> >> >>> >> ><configuration></configuration>.
> >> >>> >> >   The creation of spas worked but the Slider AM still shows the
> >> >>>same
> >> >>> >> >issue.
> >> >>> >> >   i.e. when RM1 goes from active to standby, slider AM goes
> >>from
> >> >>> >>RUNNING
> >> >>> >> >to
> >> >>> >> >   ACCPTED state with same error about TOKEN. Also NOTE that
> >>when
> >> >>> >> >   slider-client.xml is empty, the "slider destroy xxx" command
> >> >>>still
> >> >>> >> >fails
> >> >>> >> >   with Zookeeper connection errors.
> >> >>> >> >   5. I then added same parameters (as my last email - except
> >> >>> >> >   HADOOP_CONF_DIR) to slider-client.xml and ran. This time
> >> >>> >>slider-env.sh
> >> >>> >> >   has HADOOP_CONF_DIR pointing to /data/latest/conf and
> >> >>> >>slider-client.xml
> >> >>> >> >   does not have HADOOP_CONF_DIR. The same issue exists (but
> >> >>>"slider
> >> >>> >> >   destroy" does not fails)
> >> >>> >> >   6. Could you explain what do you expect to pick up from
> >>Hadoop
> >> >>> >> >   configurations that will help you in RM Token ? If slider has
> >> >>>token
> >> >>> >> >from
> >> >>> >> >   RM1, and it switches to RM2, not clear what slider does to
> >>get
> >> >>> >> >delegation
> >> >>> >> >   token for RM2 communication ?
> >> >>> >> >   7. It is worth repeating again that issue happens only when
> >>RM1
> >> >>>was
> >> >>> >> >   active when slider app was created and then RM1 becomes
> >> >>>standby. If
> >> >>> >> >RM2 was
> >> >>> >> >   active when slider app was created, then slider AM keeps
> >>running
> >> >>> for
> >> >>> >> >any
> >> >>> >> >   number of switches between RM2 and RM1 back and forth ...
> >> >>> >> >
> >> >>> >> >
> >> >>> >> >On Mon, Jul 25, 2016 at 4:21 PM, Gour Saha
> >><gs...@hortonworks.com>
> >> >>> >>wrote:
> >> >>> >> >
> >> >>> >> >> The node you are running slider from, is that a gateway node?
> >> >>>Sorry
> >> >>> >>for
> >> >>> >> >> not being explicit. I meant copy everything under
> >> >>>/etc/hadoop/conf
> >> >>> >>from
> >> >>> >> >> your cluster into some temp directory (say /tmp/hadoop_conf)
> >>in
> >> >>>your
> >> >>> >> >> gateway node or local or whichever node you are running slider
> >> >>>from.
> >> >>> >> >>Then
> >> >>> >> >> set HADOOP_CONF_DIR to /tmp/hadoop_conf and clear everything
> >>out
> >> >>> from
> >> >>> >> >> slider-client.xml.
> >> >>> >> >>
> >> >>> >> >> On 7/25/16, 4:12 PM, "Manoj Samel" <manojsamelt...@gmail.com>
> >> >>> wrote:
> >> >>> >> >>
> >> >>> >> >> >Hi Gour,
> >> >>> >> >> >
> >> >>> >> >> >Thanks for your prompt reply.
> >> >>> >> >> >
> >> >>> >> >> >FYI, issue happens when I create slider app when rm1 is
> >>active
> >> >>>and
> >> >>> >>when
> >> >>> >> >> >rm1
> >> >>> >> >> >fails over to rm2. As soon as rm2 becomes active; the slider
> >>AM
> >> >>> goes
> >> >>> >> >>from
> >> >>> >> >> >RUNNING to ACCEPTED state with above error.
> >> >>> >> >> >
> >> >>> >> >> >For your suggestion, I did following
> >> >>> >> >> >
> >> >>> >> >> >1) Copied core-site, hdfs-site, yarn-site, and mapred-site
> >>from
> >> >>> >> >> >HADOOP_CONF_DIR
> >> >>> >> >> >to slider conf directory.
> >> >>> >> >> >2) Our slider-env.sh already had HADOOP_CONF_DIR set
> >> >>> >> >> >3) I removed all properties from slider-client.xml EXCEPT
> >> >>>following
> >> >>> >> >> >
> >> >>> >> >> >   - HADOOP_CONF_DIR
> >> >>> >> >> >   - slider.yarn.queue
> >> >>> >> >> >   - slider.zookeeper.quorum
> >> >>> >> >> >   - hadoop.registry.zk.quorum
> >> >>> >> >> >   - hadoop.registry.zk.root
> >> >>> >> >> >   - hadoop.security.authorization
> >> >>> >> >> >   - hadoop.security.authentication
> >> >>> >> >> >
> >> >>> >> >> >Then I made rm1 active, installed and created slider app and
> >> >>> >>restarted
> >> >>> >> >>rm1
> >> >>> >> >> >(to make rm2) active. The slider-am again went from RUNNING
> >>to
> >> >>> >>ACCEPTED
> >> >>> >> >> >state.
> >> >>> >> >> >
> >> >>> >> >> >Let me know if you want me to try further changes.
> >> >>> >> >> >
> >> >>> >> >> >If I make the slider-client.xml completely empty per your
> >> >>> >>suggestion,
> >> >>> >> >>only
> >> >>> >> >> >slider AM comes up but it
> >> >>> >> >> >fails to start components. The AM log shows errors trying to
> >> >>> >>connect to
> >> >>> >> >> >zookeeper like below.
> >> >>> >> >> >2016-07-25 23:07:41,532
> >> >>> >> >> >[AmExecutor-006-SendThread(localhost.localdomain:2181)] WARN
> >> >>> >> >> >zookeeper.ClientCnxn - Session 0x0 for server null,
> >>unexpected
> >> >>> >>error,
> >> >>> >> >> >closing socket connection and attempting reconnect
> >> >>> >> >> >java.net.ConnectException: Connection refused
> >> >>> >> >> >
> >> >>> >> >> >Hence I kept minimal info in slider-client.xml
> >> >>> >> >> >
> >> >>> >> >> >FYI This is slider version 0.80
> >> >>> >> >> >
> >> >>> >> >> >Thanks,
> >> >>> >> >> >
> >> >>> >> >> >Manoj
> >> >>> >> >> >
> >> >>> >> >> >On Mon, Jul 25, 2016 at 2:54 PM, Gour Saha
> >> >>><gs...@hortonworks.com>
> >> >>> >> >>wrote:
> >> >>> >> >> >
> >> >>> >> >> >> If possible, can you copy the entire content of the
> >>directory
> >> >>> >> >> >> /etc/hadoop/conf and then set HADOOP_CONF_DIR in
> >> >>>slider-env.sh to
> >> >>> >>it.
> >> >>> >> >> >>Keep
> >> >>> >> >> >> slider-client.xml empty.
> >> >>> >> >> >>
> >> >>> >> >> >> Now when you do the same rm1->rm2 and then the reverse
> >> >>>failovers,
> >> >>> >>do
> >> >>> >> >>you
> >> >>> >> >> >> see the same behaviors?
> >> >>> >> >> >>
> >> >>> >> >> >> -Gour
> >> >>> >> >> >>
> >> >>> >> >> >> On 7/25/16, 2:28 PM, "Manoj Samel"
> >><manojsamelt...@gmail.com>
> >> >>> >>wrote:
> >> >>> >> >> >>
> >> >>> >> >> >> >Another observation (whatever it is worth)
> >> >>> >> >> >> >
> >> >>> >> >> >> >If slider app is created and started when rm2 was active,
> >> >>>then
> >> >>> it
> >> >>> >> >> >>seems to
> >> >>> >> >> >> >survive switches between rm2 and rm1 (and back). I.e
> >> >>> >> >> >> >
> >> >>> >> >> >> >* rm2 is active
> >> >>> >> >> >> >* create and start slider application
> >> >>> >> >> >> >* fail over to rm1. Now the Slider AM keeps running
> >> >>> >> >> >> >* fail over to rm2 again. Slider AM still keeps running
> >> >>> >> >> >> >
> >> >>> >> >> >> >So, it seems if it starts with rm1 active, then the AM
> >>goes
> >> >>>to
> >> >>> >> >> >>"ACCEPTED"
> >> >>> >> >> >> >state when RM fails to rm2. If it starts with rm2 active,
> >> >>>then
> >> >>> it
> >> >>> >> >>runs
> >> >>> >> >> >> >fine
> >> >>> >> >> >> >with any switches between rm1 and rm2.
> >> >>> >> >> >> >
> >> >>> >> >> >> >Any feedback ?
> >> >>> >> >> >> >
> >> >>> >> >> >> >Thanks,
> >> >>> >> >> >> >
> >> >>> >> >> >> >Manoj
> >> >>> >> >> >> >
> >> >>> >> >> >> >On Mon, Jul 25, 2016 at 12:25 PM, Manoj Samel
> >> >>> >> >> >><manojsamelt...@gmail.com>
> >> >>> >> >> >> >wrote:
> >> >>> >> >> >> >
> >> >>> >> >> >> >> Setup
> >> >>> >> >> >> >>
> >> >>> >> >> >> >> - Hadoop 2.6 with RM HA, Kerberos enabled
> >> >>> >> >> >> >> - Slider 0.80
> >> >>> >> >> >> >> - In my slider-client.xml, I have added all RM HA
> >> >>>properties,
> >> >>> >> >> >>including
> >> >>> >> >> >> >> the ones mentioned in
> >> >>> >> >>http://markmail.org/message/wnhpp2zn6ixo65e3.
> >> >>> >> >> >> >>
> >> >>> >> >> >> >> Following is the issue
> >> >>> >> >> >> >>
> >> >>> >> >> >> >> * rm1 is active, rm2 is standby
> >> >>> >> >> >> >> * deploy and start slider application, it runs fine
> >> >>> >> >> >> >> * restart rm1, rm2 is now active.
> >> >>> >> >> >> >> * The slider-am now goes from running into "ACCEPTED"
> >> >>>mode. It
> >> >>> >> >>stays
> >> >>> >> >> >> >>there
> >> >>> >> >> >> >> till rm1 is made active again.
> >> >>> >> >> >> >>
> >> >>> >> >> >> >> In the slider-am log, it tries to connect to RM2 and
> >> >>> connection
> >> >>> >> >>fails
> >> >>> >> >> >> >>due
> >> >>> >> >> >> >> to org.apache.hadoop.security.AccessControlException:
> >> >>>Client
> >> >>> >> >>cannot
> >> >>> >> >> >> >> authenticate via:[TOKEN]. See detailed log below
> >> >>> >> >> >> >>
> >> >>> >> >> >> >>  It seems it has some token (delegation token?) for RM1
> >>but
> >> >>> >>tries
> >> >>> >> >>to
> >> >>> >> >> >>use
> >> >>> >> >> >> >> same(?) for RM2 and fails. Am I missing some
> >>configuration
> >> >>>???
> >> >>> >> >> >> >>
> >> >>> >> >> >> >> Thanks,
> >> >>> >> >> >> >>
> >> >>> >> >> >> >>
> >> >>> >> >> >> >>
> >> >>> >> >> >> >> 2016-07-25 19:06:48,088 [AMRM Heartbeater thread] INFO
> >> >>> >> >> >> >>  client.ConfiguredRMFailoverProxyProvider - Failing
> >>over to
> >> >>> rm2
> >> >>> >> >> >> >> 2016-07-25 19:06:48,088 [AMRM Heartbeater thread] WARN
> >> >>> >> >> >> >>  security.UserGroupInformation -
> >>PriviledgedActionException
> >> >>> >> >> >>as:abc@XYZ
> >> >>> >> >> >> >> (auth:KERBEROS)
> >> >>> >> >> >>cause:org.apache.hadoop.security.AccessControlException:
> >> >>> >> >> >> >> Client cannot authenticate via:[TOKEN]
> >> >>> >> >> >> >> 2016-07-25 19:06:48,088 [AMRM Heartbeater thread] WARN
> >> >>> >> >>ipc.Client -
> >> >>> >> >> >> >> Exception encountered while connecting to the server :
> >> >>> >> >> >> >> org.apache.hadoop.security.AccessControlException:
> >>Client
> >> >>> >>cannot
> >> >>> >> >> >> >> authenticate via:[TOKEN]
> >> >>> >> >> >> >> 2016-07-25 19:06:48,088 [AMRM Heartbeater thread] WARN
> >> >>> >> >> >> >>  security.UserGroupInformation -
> >>PriviledgedActionException
> >> >>> >> >> >>as:abc@XYZ
> >> >>> >> >> >> >> (auth:KERBEROS) cause:java.io.IOException:
> >> >>> >> >> >> >> org.apache.hadoop.security.AccessControlException:
> >>Client
> >> >>> >>cannot
> >> >>> >> >> >> >> authenticate via:[TOKEN]
> >> >>> >> >> >> >> 2016-07-25 19:06:48,089 [AMRM Heartbeater thread] INFO
> >> >>> >> >> >> >>  retry.RetryInvocationHandler - Exception while invoking
> >> >>> >>allocate
> >> >>> >> >>of
> >> >>> >> >> >> >>class
> >> >>> >> >> >> >> ApplicationMasterProtocolPBClientImpl over rm2 after 287
> >> >>>fail
> >> >>> >>over
> >> >>> >> >> >> >> attempts. Trying to fail over immediately.
> >> >>> >> >> >> >> java.io.IOException: Failed on local exception:
> >> >>> >> >>java.io.IOException:
> >> >>> >> >> >> >> org.apache.hadoop.security.AccessControlException:
> >>Client
> >> >>> >>cannot
> >> >>> >> >> >> >> authenticate via:[TOKEN]; Host Details : local host is:
> >> >>> >>"<SliderAM
> >> >>> >> >> >> >> HOST>/<slider AM Host IP>"; destination host is: "<RM2
> >> >>> >> >>HOST>":23130;
> >> >>> >> >> >> >>         at
> >> >>> >> >> >>
> >> >>>>>org.apache.hadoop.net.NetUtils.wrapException(NetUtils.java:772)
> >> >>> >> >> >> >>         at
> >> >>>org.apache.hadoop.ipc.Client.call(Client.java:1476)
> >> >>> >> >> >> >>         at
> >> >>>org.apache.hadoop.ipc.Client.call(Client.java:1403)
> >> >>> >> >> >> >>         at
> >> >>> >> >> >> >>
> >> >>> >> >> >>
> >> >>> >> >>
> >> >>> >>
> >> >>>
> >> >>>
> >>
> >>>>>>>>>>>>>org.apache.hadoop.ipc.ProtobufRpcEngine$Invoker.invoke(Protobu
> >>>>>>>>>>>>>fR
> >> >>>>>>>>>>>pcE
> >> >>> >>>>>>>>ng
> >> >>> >> >>>>>>in
> >> >>> >> >> >>>>e.
> >> >>> >> >> >> >>java:230)
> >> >>> >> >> >> >>         at com.sun.proxy.$Proxy23.allocate(Unknown
> >>Source)
> >> >>> >> >> >> >>         at
> >> >>> >> >> >> >>
> >> >>> >> >> >>
> >> >>> >> >>
> >> >>> >>
> >> >>>
> >> >>>
> >>
> >>>>>>>>>>>>>org.apache.hadoop.yarn.api.impl.pb.client.ApplicationMasterPro
> >>>>>>>>>>>>>to
> >> >>>>>>>>>>>col
> >> >>> >>>>>>>>PB
> >> >>> >> >>>>>>Cl
> >> >>> >> >> >>>>ie
> >> >>> >> >> >>
> >> >>>>>ntImpl.allocate(ApplicationMasterProtocolPBClientImpl.java:77)
> >> >>> >> >> >> >>         at
> >> >>> sun.reflect.GeneratedMethodAccessor10.invoke(Unknown
> >> >>> >> >> >>Source)
> >> >>> >> >> >> >>         at
> >> >>> >> >> >> >>
> >> >>> >> >> >>
> >> >>> >> >>
> >> >>> >>
> >> >>>
> >> >>>
> >>
> >>>>>>>>>>>>>sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMeth
> >>>>>>>>>>>>>od
> >> >>>>>>>>>>>Acc
> >> >>> >>>>>>>>es
> >> >>> >> >>>>>>so
> >> >>> >> >> >>>>rI
> >> >>> >> >> >> >>mpl.java:43)
> >> >>> >> >> >> >>         at
> >>java.lang.reflect.Method.invoke(Method.java:497)
> >> >>> >> >> >> >>         at
> >> >>> >> >> >> >>
> >> >>> >> >> >>
> >> >>> >> >>
> >> >>> >>
> >> >>>
> >> >>>
> >>
> >>>>>>>>>>>>>org.apache.hadoop.io.retry.RetryInvocationHandler.invokeMethod
> >>>>>>>>>>>>>(R
> >> >>>>>>>>>>>etr
> >> >>> >>>>>>>>yI
> >> >>> >> >>>>>>nv
> >> >>> >> >> >>>>oc
> >> >>> >> >> >> >>ationHandler.java:252)
> >> >>> >> >> >> >>         at
> >> >>> >> >> >> >>
> >> >>> >> >> >>
> >> >>> >> >>
> >> >>> >>
> >> >>>
> >> >>>
> >>
> >>>>>>>>>>>>>org.apache.hadoop.io.retry.RetryInvocationHandler.invoke(Retry
> >>>>>>>>>>>>>In
> >> >>>>>>>>>>>voc
> >> >>> >>>>>>>>at
> >> >>> >> >>>>>>io
> >> >>> >> >> >>>>nH
> >> >>> >> >> >> >>andler.java:104)
> >> >>> >> >> >> >>         at com.sun.proxy.$Proxy24.allocate(Unknown
> >>Source)
> >> >>> >> >> >> >>         at
> >> >>> >> >> >> >>
> >> >>> >> >> >>
> >> >>> >> >>
> >> >>> >>
> >> >>>
> >> >>>
> >>
> >>>>>>>>>>>>>org.apache.hadoop.yarn.client.api.impl.AMRMClientImpl.allocate
> >>>>>>>>>>>>>(A
> >> >>>>>>>>>>>MRM
> >> >>> >>>>>>>>Cl
> >> >>> >> >>>>>>ie
> >> >>> >> >> >>>>nt
> >> >>> >> >> >> >>Impl.java:278)
> >> >>> >> >> >> >>         at
> >> >>> >> >> >> >>
> >> >>> >> >> >>
> >> >>> >> >>
> >> >>> >>
> >> >>>
> >> >>>
> >>
> >>>>>>>>>>>>>org.apache.hadoop.yarn.client.api.async.impl.AMRMClientAsyncIm
> >>>>>>>>>>>>>pl
> >> >>>>>>>>>>>$He
> >> >>> >>>>>>>>ar
> >> >>> >> >>>>>>tb
> >> >>> >> >> >>>>ea
> >> >>> >> >> >> >>tThread.run(AMRMClientAsyncImpl.java:224)
> >> >>> >> >> >> >> Caused by: java.io.IOException:
> >> >>> >> >> >> >> org.apache.hadoop.security.AccessControlException:
> >>Client
> >> >>> >>cannot
> >> >>> >> >> >> >> authenticate via:[TOKEN]
> >> >>> >> >> >> >>         at
> >> >>> >> >> >>
> >> >>>>>org.apache.hadoop.ipc.Client$Connection$1.run(Client.java:682)
> >> >>> >> >> >> >>         at
> >> >>>java.security.AccessController.doPrivileged(Native
> >> >>> >> >>Method)
> >> >>> >> >> >> >>         at
> >> >>>javax.security.auth.Subject.doAs(Subject.java:422)
> >> >>> >> >> >> >>         at
> >> >>> >> >> >> >>
> >> >>> >> >> >>
> >> >>> >> >>
> >> >>> >>
> >> >>>
> >> >>>
> >>
> >>>>>>>>>>>>>org.apache.hadoop.security.UserGroupInformation.doAs(UserGroup
> >>>>>>>>>>>>>In
> >> >>>>>>>>>>>for
> >> >>> >>>>>>>>ma
> >> >>> >> >>>>>>ti
> >> >>> >> >> >>>>on
> >> >>> >> >> >> >>.java:1671)
> >> >>> >> >> >> >>         at
> >> >>> >> >> >> >>
> >> >>> >> >> >>
> >> >>> >> >>
> >> >>> >>
> >> >>>
> >> >>>
> >>
> >>>>>>>>>>>>>org.apache.hadoop.ipc.Client$Connection.handleSaslConnectionFa
> >>>>>>>>>>>>>il
> >> >>>>>>>>>>>ure
> >> >>> >>>>>>>>(C
> >> >>> >> >>>>>>li
> >> >>> >> >> >>>>en
> >> >>> >> >> >> >>t.java:645)
> >> >>> >> >> >> >>         at
> >> >>> >> >> >> >>
> >> >>> >> >>
> >> >>> >>
> >> >>>
> >> >>>
> >>
> >>>>>>>>>>>org.apache.hadoop.ipc.Client$Connection.setupIOstreams(Client.ja
> >>>>>>>>>>>va
> >> >>>>>>>>>:73
> >> >>> >>>>>>3)
> >> >>> >> >> >> >>         at
> >> >>> >> >> >> >>
> >> >>> >>
> >>
> >>>>>>>org.apache.hadoop.ipc.Client$Connection.access$2800(Client.java:370)
> >> >>> >> >> >> >>         at
> >> >>> >> >> >>org.apache.hadoop.ipc.Client.getConnection(Client.java:1525)
> >> >>> >> >> >> >>         at
> >> >>>org.apache.hadoop.ipc.Client.call(Client.java:1442)
> >> >>> >> >> >> >>         ... 12 more
> >> >>> >> >> >> >> Caused by:
> >> >>>org.apache.hadoop.security.AccessControlException:
> >> >>> >> >>Client
> >> >>> >> >> >> >> cannot authenticate via:[TOKEN]
> >> >>> >> >> >> >>         at
> >> >>> >> >> >> >>
> >> >>> >> >> >>
> >> >>> >> >>
> >> >>> >>
> >> >>>
> >> >>>
> >>
> >>>>>>>>>>>>>org.apache.hadoop.security.SaslRpcClient.selectSaslClient(Sasl
> >>>>>>>>>>>>>Rp
> >> >>>>>>>>>>>cCl
> >> >>> >>>>>>>>ie
> >> >>> >> >>>>>>nt
> >> >>> >> >> >>>>.j
> >> >>> >> >> >> >>ava:172)
> >> >>> >> >> >> >>         at
> >> >>> >> >> >> >>
> >> >>> >> >> >>
> >> >>> >> >>
> >> >>> >>
> >> >>>
> >> >>>
> >>
> >>>>>>>>>>>>>org.apache.hadoop.security.SaslRpcClient.saslConnect(SaslRpcCl
> >>>>>>>>>>>>>ie
> >> >>>>>>>>>>>nt.
> >> >>> >>>>>>>>ja
> >> >>> >> >>>>>>va
> >> >>> >> >> >>>>:3
> >> >>> >> >> >> >>96)
> >> >>> >> >> >> >>         at
> >> >>> >> >> >> >>
> >> >>> >> >> >>
> >> >>> >> >>
> >> >>> >>
> >> >>>
> >> >>>
> >>
> >>>>>>>>>>>>>org.apache.hadoop.ipc.Client$Connection.setupSaslConnection(Cl
> >>>>>>>>>>>>>ie
> >> >>>>>>>>>>>nt.
> >> >>> >>>>>>>>ja
> >> >>> >> >>>>>>va
> >> >>> >> >> >>>>:5
> >> >>> >> >> >> >>55)
> >> >>> >> >> >> >>         at
> >> >>> >> >> >> >>
> >> >>> >>
> >>
> >>>>>>>org.apache.hadoop.ipc.Client$Connection.access$1800(Client.java:370)
> >> >>> >> >> >> >>         at
> >> >>> >> >> >>
> >> >>>>>org.apache.hadoop.ipc.Client$Connection$2.run(Client.java:725)
> >> >>> >> >> >> >>         at
> >> >>> >> >> >>
> >> >>>>>org.apache.hadoop.ipc.Client$Connection$2.run(Client.java:721)
> >> >>> >> >> >> >>         at
> >> >>>java.security.AccessController.doPrivileged(Native
> >> >>> >> >>Method)
> >> >>> >> >> >> >>         at
> >> >>>javax.security.auth.Subject.doAs(Subject.java:422)
> >> >>> >> >> >> >>         at
> >> >>> >> >> >> >>
> >> >>> >> >> >>
> >> >>> >> >>
> >> >>> >>
> >> >>>
> >> >>>
> >>
> >>>>>>>>>>>>>org.apache.hadoop.security.UserGroupInformation.doAs(UserGroup
> >>>>>>>>>>>>>In
> >> >>>>>>>>>>>for
> >> >>> >>>>>>>>ma
> >> >>> >> >>>>>>ti
> >> >>> >> >> >>>>on
> >> >>> >> >> >> >>.java:1671)
> >> >>> >> >> >> >>         at
> >> >>> >> >> >> >>
> >> >>> >> >>
> >> >>> >>
> >> >>>
> >> >>>
> >>
> >>>>>>>>>>>org.apache.hadoop.ipc.Client$Connection.setupIOstreams(Client.ja
> >>>>>>>>>>>va
> >> >>>>>>>>>:72
> >> >>> >>>>>>0)
> >> >>> >> >> >> >>         ... 15 more
> >> >>> >> >> >> >> 2016-07-25 19:06:48,089 [AMRM Heartbeater thread] INFO
> >> >>> >> >> >> >>  client.ConfiguredRMFailoverProxyProvider - Failing
> >>over to
> >> >>> rm1
> >> >>> >> >> >> >>
> >> >>> >> >> >>
> >> >>> >> >> >>
> >> >>> >> >>
> >> >>> >> >>
> >> >>> >>
> >> >>> >>
> >> >>>
> >> >>>
> >> >>
> >>
> >>
>
>

Reply via email to