That is strange, since it is indeed not required to contain anything in
slider-client.xml (except <configuration></configuration>) if
HADOOP_CONF_DIR has everything that Slider needs. This probably gives an
indication that there might be some issue with cluster configuration based
on files solely under HADOOP_CONF_DIR to begin with.

Suggest you to upload all the config files to the jira to help debug this
further.

-Gour

On 7/28/16, 4:27 PM, "Manoj Samel" <manojsamelt...@gmail.com> wrote:

>Thanks Gour for prompt reply
>
>BTW - Creating a empty slider-client.xml (with just
><configuration></configuration>) does not works. The AM starts but fails
>to
>create any components and shows errors like
>
>2016-07-28 23:18:46,018
>[AmExecutor-006-SendThread(localhost.localdomain:2181)] WARN
> zookeeper.ClientCnxn - Session 0x0 for server null, unexpected error,
>closing socket connection and attempting reconnect
>java.net.ConnectException: Connection refused
>        at sun.nio.ch.SocketChannelImpl.checkConnect(Native Method)
>        at
>sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:717)
>        at
>org.apache.zookeeper.ClientCnxnSocketNIO.doTransport(ClientCnxnSocketNIO.j
>ava:361)
>        at
>org.apache.zookeeper.ClientCnxn$SendThread.run(ClientCnxn.java:1081)
>
>Also, command "slider destroy <app>" fails with zookeeper errors ...
>
>I had to keep a minimal slider-client.xml. It does not have any RM info
>etc. but does contain slider ZK related properties like
>"slider.zookeeper.quorum", "hadoop.registry.zk.quorum",
>"hadoop.registry.zk.root". I haven't yet distilled the absolute minimal
>set
>of properties required, but this should suffice for now. All RM / HDFS
>properties will be read from HADOOP_CONF_DIR files.
>
>Let me know if this could cause any issues.
>
>On Thu, Jul 28, 2016 at 3:36 PM, Gour Saha <gs...@hortonworks.com> wrote:
>
>> No need to copy any files. Pointing HADOOP_CONF_DIR to /etc/hadoop/conf
>>is
>> good.
>>
>> -Gour
>>
>> On 7/28/16, 3:24 PM, "Manoj Samel" <manojsamelt...@gmail.com> wrote:
>>
>> >Follow up question regarding Gour's comment in earlier thread -
>> >
>> >Slider is installed on one of the hadoop nodes. SLIDER_HOME/conf
>>directory
>> >(say /data/slider/conf) is different than HADOOP_CONF_DIR
>> >(/etc/hadoop/conf). Is it required/recommended that files in
>> >HADOOP_CONF_DIR be copied to SLIDER_HOME/conf and slider-env.sh script
>> >sets
>> >HADOOP_CONF_DIR to /data/slider/conf ?
>> >
>> >Or can the slider-env.sh set HADOOP_CONF_DIR to /etc/hadoop/conf ,
>>without
>> >copying the files ?
>> >
>> >Using slider .80 for now, but would like to know recommendation for
>>this
>> >and future versions as well.
>> >
>> >Thanks in advance,
>> >
>> >Manoj
>> >
>> >On Tue, Jul 26, 2016 at 3:27 PM, Manoj Samel <manojsamelt...@gmail.com>
>> >wrote:
>> >
>> >> Filed https://issues.apache.org/jira/browse/SLIDER-1158 with logs and
>> my
>> >> analysis of logs.
>> >>
>> >> On Tue, Jul 26, 2016 at 10:36 AM, Gour Saha <gs...@hortonworks.com>
>> >>wrote:
>> >>
>> >>> Please file a JIRA and upload the logs to it.
>> >>>
>> >>> On 7/26/16, 10:21 AM, "Manoj Samel" <manojsamelt...@gmail.com>
>>wrote:
>> >>>
>> >>> >Hi Gour,
>> >>> >
>> >>> >Can you please reach me using your own email-id? I will then send
>> >>>logs to
>> >>> >you, along with my analysis - I don't want to send logs on public
>>list
>> >>> >
>> >>> >Thanks,
>> >>> >
>> >>> >On Mon, Jul 25, 2016 at 5:39 PM, Gour Saha <gs...@hortonworks.com>
>> >>> wrote:
>> >>> >
>> >>> >> Ok, so this node is not a gateway. It is part of the cluster,
>>which
>> >>> >>means
>> >>> >> you don¹t need slider-client.xml at all. Just have
>>HADOOP_CONF_DIR
>> >>> >> pointing to /etc/hadoop/conf in slider-env.sh and that should be
>>it.
>> >>> >>
>> >>> >> So the above simplifies your config setup. It will not solve
>>either
>> >>>of
>> >>> >>the
>> >>> >> 2 problems you are facing.
>> >>> >>
>> >>> >> Now coming to the 2 issues you are facing, you have to provide
>> >>> >>additional
>> >>> >> logs for us to understand better. Let¹s start with  -
>> >>> >> 1. RM logs (specifically between the time when rm1->rm2 failover
>>is
>> >>> >> simulated)
>> >>> >> 2. Slider App logs
>> >>> >>
>> >>> >> -Gour
>> >>> >>
>> >>> >> On 7/25/16, 5:16 PM, "Manoj Samel" <manojsamelt...@gmail.com>
>> wrote:
>> >>> >>
>> >>> >> >   1. Not clear about your question on "gateway" node. The node
>> >>> running
>> >>> >> >   slider is part of the hadoop cluster and there are other
>> >>>services
>> >>> >>like
>> >>> >> >   Oozie that run on this node that utilizes hdfs and yarn. So
>>if
>> >>>your
>> >>> >> >   question is whether the node is otherwise working for HDFS
>>and
>> >>>Yarn
>> >>> >> >   configuration, it is working
>> >>> >> >   2. I copied all files from HADOOP_CONF_DIR (say
>> >>>/etc/hadoop/conf)
>> >>> to
>> >>> >> >the
>> >>> >> >   directory containing slider-client.xml (say
>>/data/latest/conf)
>> >>> >> >   3. In earlier email, I had done a mistake where slider-env.sh
>> >>>file
>> >>> >> >HADOOP_CONF_DIR
>> >>> >> >   was pointing to original directory /etc/hadoop/conf. I edited
>> >>>it to
>> >>> >> >   point to same directory containing slider-client.xml &
>> >>> slider-env.sh
>> >>> >> >i.e.
>> >>> >> >   /data/latest/conf
>> >>> >> >   4. I emptied slider-client.xml. It just had the
>> >>> >> ><configuration></configuration>.
>> >>> >> >   The creation of spas worked but the Slider AM still shows the
>> >>>same
>> >>> >> >issue.
>> >>> >> >   i.e. when RM1 goes from active to standby, slider AM goes
>>from
>> >>> >>RUNNING
>> >>> >> >to
>> >>> >> >   ACCPTED state with same error about TOKEN. Also NOTE that
>>when
>> >>> >> >   slider-client.xml is empty, the "slider destroy xxx" command
>> >>>still
>> >>> >> >fails
>> >>> >> >   with Zookeeper connection errors.
>> >>> >> >   5. I then added same parameters (as my last email - except
>> >>> >> >   HADOOP_CONF_DIR) to slider-client.xml and ran. This time
>> >>> >>slider-env.sh
>> >>> >> >   has HADOOP_CONF_DIR pointing to /data/latest/conf and
>> >>> >>slider-client.xml
>> >>> >> >   does not have HADOOP_CONF_DIR. The same issue exists (but
>> >>>"slider
>> >>> >> >   destroy" does not fails)
>> >>> >> >   6. Could you explain what do you expect to pick up from
>>Hadoop
>> >>> >> >   configurations that will help you in RM Token ? If slider has
>> >>>token
>> >>> >> >from
>> >>> >> >   RM1, and it switches to RM2, not clear what slider does to
>>get
>> >>> >> >delegation
>> >>> >> >   token for RM2 communication ?
>> >>> >> >   7. It is worth repeating again that issue happens only when
>>RM1
>> >>>was
>> >>> >> >   active when slider app was created and then RM1 becomes
>> >>>standby. If
>> >>> >> >RM2 was
>> >>> >> >   active when slider app was created, then slider AM keeps
>>running
>> >>> for
>> >>> >> >any
>> >>> >> >   number of switches between RM2 and RM1 back and forth ...
>> >>> >> >
>> >>> >> >
>> >>> >> >On Mon, Jul 25, 2016 at 4:21 PM, Gour Saha
>><gs...@hortonworks.com>
>> >>> >>wrote:
>> >>> >> >
>> >>> >> >> The node you are running slider from, is that a gateway node?
>> >>>Sorry
>> >>> >>for
>> >>> >> >> not being explicit. I meant copy everything under
>> >>>/etc/hadoop/conf
>> >>> >>from
>> >>> >> >> your cluster into some temp directory (say /tmp/hadoop_conf)
>>in
>> >>>your
>> >>> >> >> gateway node or local or whichever node you are running slider
>> >>>from.
>> >>> >> >>Then
>> >>> >> >> set HADOOP_CONF_DIR to /tmp/hadoop_conf and clear everything
>>out
>> >>> from
>> >>> >> >> slider-client.xml.
>> >>> >> >>
>> >>> >> >> On 7/25/16, 4:12 PM, "Manoj Samel" <manojsamelt...@gmail.com>
>> >>> wrote:
>> >>> >> >>
>> >>> >> >> >Hi Gour,
>> >>> >> >> >
>> >>> >> >> >Thanks for your prompt reply.
>> >>> >> >> >
>> >>> >> >> >FYI, issue happens when I create slider app when rm1 is
>>active
>> >>>and
>> >>> >>when
>> >>> >> >> >rm1
>> >>> >> >> >fails over to rm2. As soon as rm2 becomes active; the slider
>>AM
>> >>> goes
>> >>> >> >>from
>> >>> >> >> >RUNNING to ACCEPTED state with above error.
>> >>> >> >> >
>> >>> >> >> >For your suggestion, I did following
>> >>> >> >> >
>> >>> >> >> >1) Copied core-site, hdfs-site, yarn-site, and mapred-site
>>from
>> >>> >> >> >HADOOP_CONF_DIR
>> >>> >> >> >to slider conf directory.
>> >>> >> >> >2) Our slider-env.sh already had HADOOP_CONF_DIR set
>> >>> >> >> >3) I removed all properties from slider-client.xml EXCEPT
>> >>>following
>> >>> >> >> >
>> >>> >> >> >   - HADOOP_CONF_DIR
>> >>> >> >> >   - slider.yarn.queue
>> >>> >> >> >   - slider.zookeeper.quorum
>> >>> >> >> >   - hadoop.registry.zk.quorum
>> >>> >> >> >   - hadoop.registry.zk.root
>> >>> >> >> >   - hadoop.security.authorization
>> >>> >> >> >   - hadoop.security.authentication
>> >>> >> >> >
>> >>> >> >> >Then I made rm1 active, installed and created slider app and
>> >>> >>restarted
>> >>> >> >>rm1
>> >>> >> >> >(to make rm2) active. The slider-am again went from RUNNING
>>to
>> >>> >>ACCEPTED
>> >>> >> >> >state.
>> >>> >> >> >
>> >>> >> >> >Let me know if you want me to try further changes.
>> >>> >> >> >
>> >>> >> >> >If I make the slider-client.xml completely empty per your
>> >>> >>suggestion,
>> >>> >> >>only
>> >>> >> >> >slider AM comes up but it
>> >>> >> >> >fails to start components. The AM log shows errors trying to
>> >>> >>connect to
>> >>> >> >> >zookeeper like below.
>> >>> >> >> >2016-07-25 23:07:41,532
>> >>> >> >> >[AmExecutor-006-SendThread(localhost.localdomain:2181)] WARN
>> >>> >> >> >zookeeper.ClientCnxn - Session 0x0 for server null,
>>unexpected
>> >>> >>error,
>> >>> >> >> >closing socket connection and attempting reconnect
>> >>> >> >> >java.net.ConnectException: Connection refused
>> >>> >> >> >
>> >>> >> >> >Hence I kept minimal info in slider-client.xml
>> >>> >> >> >
>> >>> >> >> >FYI This is slider version 0.80
>> >>> >> >> >
>> >>> >> >> >Thanks,
>> >>> >> >> >
>> >>> >> >> >Manoj
>> >>> >> >> >
>> >>> >> >> >On Mon, Jul 25, 2016 at 2:54 PM, Gour Saha
>> >>><gs...@hortonworks.com>
>> >>> >> >>wrote:
>> >>> >> >> >
>> >>> >> >> >> If possible, can you copy the entire content of the
>>directory
>> >>> >> >> >> /etc/hadoop/conf and then set HADOOP_CONF_DIR in
>> >>>slider-env.sh to
>> >>> >>it.
>> >>> >> >> >>Keep
>> >>> >> >> >> slider-client.xml empty.
>> >>> >> >> >>
>> >>> >> >> >> Now when you do the same rm1->rm2 and then the reverse
>> >>>failovers,
>> >>> >>do
>> >>> >> >>you
>> >>> >> >> >> see the same behaviors?
>> >>> >> >> >>
>> >>> >> >> >> -Gour
>> >>> >> >> >>
>> >>> >> >> >> On 7/25/16, 2:28 PM, "Manoj Samel"
>><manojsamelt...@gmail.com>
>> >>> >>wrote:
>> >>> >> >> >>
>> >>> >> >> >> >Another observation (whatever it is worth)
>> >>> >> >> >> >
>> >>> >> >> >> >If slider app is created and started when rm2 was active,
>> >>>then
>> >>> it
>> >>> >> >> >>seems to
>> >>> >> >> >> >survive switches between rm2 and rm1 (and back). I.e
>> >>> >> >> >> >
>> >>> >> >> >> >* rm2 is active
>> >>> >> >> >> >* create and start slider application
>> >>> >> >> >> >* fail over to rm1. Now the Slider AM keeps running
>> >>> >> >> >> >* fail over to rm2 again. Slider AM still keeps running
>> >>> >> >> >> >
>> >>> >> >> >> >So, it seems if it starts with rm1 active, then the AM
>>goes
>> >>>to
>> >>> >> >> >>"ACCEPTED"
>> >>> >> >> >> >state when RM fails to rm2. If it starts with rm2 active,
>> >>>then
>> >>> it
>> >>> >> >>runs
>> >>> >> >> >> >fine
>> >>> >> >> >> >with any switches between rm1 and rm2.
>> >>> >> >> >> >
>> >>> >> >> >> >Any feedback ?
>> >>> >> >> >> >
>> >>> >> >> >> >Thanks,
>> >>> >> >> >> >
>> >>> >> >> >> >Manoj
>> >>> >> >> >> >
>> >>> >> >> >> >On Mon, Jul 25, 2016 at 12:25 PM, Manoj Samel
>> >>> >> >> >><manojsamelt...@gmail.com>
>> >>> >> >> >> >wrote:
>> >>> >> >> >> >
>> >>> >> >> >> >> Setup
>> >>> >> >> >> >>
>> >>> >> >> >> >> - Hadoop 2.6 with RM HA, Kerberos enabled
>> >>> >> >> >> >> - Slider 0.80
>> >>> >> >> >> >> - In my slider-client.xml, I have added all RM HA
>> >>>properties,
>> >>> >> >> >>including
>> >>> >> >> >> >> the ones mentioned in
>> >>> >> >>http://markmail.org/message/wnhpp2zn6ixo65e3.
>> >>> >> >> >> >>
>> >>> >> >> >> >> Following is the issue
>> >>> >> >> >> >>
>> >>> >> >> >> >> * rm1 is active, rm2 is standby
>> >>> >> >> >> >> * deploy and start slider application, it runs fine
>> >>> >> >> >> >> * restart rm1, rm2 is now active.
>> >>> >> >> >> >> * The slider-am now goes from running into "ACCEPTED"
>> >>>mode. It
>> >>> >> >>stays
>> >>> >> >> >> >>there
>> >>> >> >> >> >> till rm1 is made active again.
>> >>> >> >> >> >>
>> >>> >> >> >> >> In the slider-am log, it tries to connect to RM2 and
>> >>> connection
>> >>> >> >>fails
>> >>> >> >> >> >>due
>> >>> >> >> >> >> to org.apache.hadoop.security.AccessControlException:
>> >>>Client
>> >>> >> >>cannot
>> >>> >> >> >> >> authenticate via:[TOKEN]. See detailed log below
>> >>> >> >> >> >>
>> >>> >> >> >> >>  It seems it has some token (delegation token?) for RM1
>>but
>> >>> >>tries
>> >>> >> >>to
>> >>> >> >> >>use
>> >>> >> >> >> >> same(?) for RM2 and fails. Am I missing some
>>configuration
>> >>>???
>> >>> >> >> >> >>
>> >>> >> >> >> >> Thanks,
>> >>> >> >> >> >>
>> >>> >> >> >> >>
>> >>> >> >> >> >>
>> >>> >> >> >> >> 2016-07-25 19:06:48,088 [AMRM Heartbeater thread] INFO
>> >>> >> >> >> >>  client.ConfiguredRMFailoverProxyProvider - Failing
>>over to
>> >>> rm2
>> >>> >> >> >> >> 2016-07-25 19:06:48,088 [AMRM Heartbeater thread] WARN
>> >>> >> >> >> >>  security.UserGroupInformation -
>>PriviledgedActionException
>> >>> >> >> >>as:abc@XYZ
>> >>> >> >> >> >> (auth:KERBEROS)
>> >>> >> >> >>cause:org.apache.hadoop.security.AccessControlException:
>> >>> >> >> >> >> Client cannot authenticate via:[TOKEN]
>> >>> >> >> >> >> 2016-07-25 19:06:48,088 [AMRM Heartbeater thread] WARN
>> >>> >> >>ipc.Client -
>> >>> >> >> >> >> Exception encountered while connecting to the server :
>> >>> >> >> >> >> org.apache.hadoop.security.AccessControlException:
>>Client
>> >>> >>cannot
>> >>> >> >> >> >> authenticate via:[TOKEN]
>> >>> >> >> >> >> 2016-07-25 19:06:48,088 [AMRM Heartbeater thread] WARN
>> >>> >> >> >> >>  security.UserGroupInformation -
>>PriviledgedActionException
>> >>> >> >> >>as:abc@XYZ
>> >>> >> >> >> >> (auth:KERBEROS) cause:java.io.IOException:
>> >>> >> >> >> >> org.apache.hadoop.security.AccessControlException:
>>Client
>> >>> >>cannot
>> >>> >> >> >> >> authenticate via:[TOKEN]
>> >>> >> >> >> >> 2016-07-25 19:06:48,089 [AMRM Heartbeater thread] INFO
>> >>> >> >> >> >>  retry.RetryInvocationHandler - Exception while invoking
>> >>> >>allocate
>> >>> >> >>of
>> >>> >> >> >> >>class
>> >>> >> >> >> >> ApplicationMasterProtocolPBClientImpl over rm2 after 287
>> >>>fail
>> >>> >>over
>> >>> >> >> >> >> attempts. Trying to fail over immediately.
>> >>> >> >> >> >> java.io.IOException: Failed on local exception:
>> >>> >> >>java.io.IOException:
>> >>> >> >> >> >> org.apache.hadoop.security.AccessControlException:
>>Client
>> >>> >>cannot
>> >>> >> >> >> >> authenticate via:[TOKEN]; Host Details : local host is:
>> >>> >>"<SliderAM
>> >>> >> >> >> >> HOST>/<slider AM Host IP>"; destination host is: "<RM2
>> >>> >> >>HOST>":23130;
>> >>> >> >> >> >>         at
>> >>> >> >> >>
>> >>>>>org.apache.hadoop.net.NetUtils.wrapException(NetUtils.java:772)
>> >>> >> >> >> >>         at
>> >>>org.apache.hadoop.ipc.Client.call(Client.java:1476)
>> >>> >> >> >> >>         at
>> >>>org.apache.hadoop.ipc.Client.call(Client.java:1403)
>> >>> >> >> >> >>         at
>> >>> >> >> >> >>
>> >>> >> >> >>
>> >>> >> >>
>> >>> >>
>> >>>
>> >>>
>> 
>>>>>>>>>>>>>org.apache.hadoop.ipc.ProtobufRpcEngine$Invoker.invoke(Protobu
>>>>>>>>>>>>>fR
>> >>>>>>>>>>>pcE
>> >>> >>>>>>>>ng
>> >>> >> >>>>>>in
>> >>> >> >> >>>>e.
>> >>> >> >> >> >>java:230)
>> >>> >> >> >> >>         at com.sun.proxy.$Proxy23.allocate(Unknown
>>Source)
>> >>> >> >> >> >>         at
>> >>> >> >> >> >>
>> >>> >> >> >>
>> >>> >> >>
>> >>> >>
>> >>>
>> >>>
>> 
>>>>>>>>>>>>>org.apache.hadoop.yarn.api.impl.pb.client.ApplicationMasterPro
>>>>>>>>>>>>>to
>> >>>>>>>>>>>col
>> >>> >>>>>>>>PB
>> >>> >> >>>>>>Cl
>> >>> >> >> >>>>ie
>> >>> >> >> >>
>> >>>>>ntImpl.allocate(ApplicationMasterProtocolPBClientImpl.java:77)
>> >>> >> >> >> >>         at
>> >>> sun.reflect.GeneratedMethodAccessor10.invoke(Unknown
>> >>> >> >> >>Source)
>> >>> >> >> >> >>         at
>> >>> >> >> >> >>
>> >>> >> >> >>
>> >>> >> >>
>> >>> >>
>> >>>
>> >>>
>> 
>>>>>>>>>>>>>sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMeth
>>>>>>>>>>>>>od
>> >>>>>>>>>>>Acc
>> >>> >>>>>>>>es
>> >>> >> >>>>>>so
>> >>> >> >> >>>>rI
>> >>> >> >> >> >>mpl.java:43)
>> >>> >> >> >> >>         at
>>java.lang.reflect.Method.invoke(Method.java:497)
>> >>> >> >> >> >>         at
>> >>> >> >> >> >>
>> >>> >> >> >>
>> >>> >> >>
>> >>> >>
>> >>>
>> >>>
>> 
>>>>>>>>>>>>>org.apache.hadoop.io.retry.RetryInvocationHandler.invokeMethod
>>>>>>>>>>>>>(R
>> >>>>>>>>>>>etr
>> >>> >>>>>>>>yI
>> >>> >> >>>>>>nv
>> >>> >> >> >>>>oc
>> >>> >> >> >> >>ationHandler.java:252)
>> >>> >> >> >> >>         at
>> >>> >> >> >> >>
>> >>> >> >> >>
>> >>> >> >>
>> >>> >>
>> >>>
>> >>>
>> 
>>>>>>>>>>>>>org.apache.hadoop.io.retry.RetryInvocationHandler.invoke(Retry
>>>>>>>>>>>>>In
>> >>>>>>>>>>>voc
>> >>> >>>>>>>>at
>> >>> >> >>>>>>io
>> >>> >> >> >>>>nH
>> >>> >> >> >> >>andler.java:104)
>> >>> >> >> >> >>         at com.sun.proxy.$Proxy24.allocate(Unknown
>>Source)
>> >>> >> >> >> >>         at
>> >>> >> >> >> >>
>> >>> >> >> >>
>> >>> >> >>
>> >>> >>
>> >>>
>> >>>
>> 
>>>>>>>>>>>>>org.apache.hadoop.yarn.client.api.impl.AMRMClientImpl.allocate
>>>>>>>>>>>>>(A
>> >>>>>>>>>>>MRM
>> >>> >>>>>>>>Cl
>> >>> >> >>>>>>ie
>> >>> >> >> >>>>nt
>> >>> >> >> >> >>Impl.java:278)
>> >>> >> >> >> >>         at
>> >>> >> >> >> >>
>> >>> >> >> >>
>> >>> >> >>
>> >>> >>
>> >>>
>> >>>
>> 
>>>>>>>>>>>>>org.apache.hadoop.yarn.client.api.async.impl.AMRMClientAsyncIm
>>>>>>>>>>>>>pl
>> >>>>>>>>>>>$He
>> >>> >>>>>>>>ar
>> >>> >> >>>>>>tb
>> >>> >> >> >>>>ea
>> >>> >> >> >> >>tThread.run(AMRMClientAsyncImpl.java:224)
>> >>> >> >> >> >> Caused by: java.io.IOException:
>> >>> >> >> >> >> org.apache.hadoop.security.AccessControlException:
>>Client
>> >>> >>cannot
>> >>> >> >> >> >> authenticate via:[TOKEN]
>> >>> >> >> >> >>         at
>> >>> >> >> >>
>> >>>>>org.apache.hadoop.ipc.Client$Connection$1.run(Client.java:682)
>> >>> >> >> >> >>         at
>> >>>java.security.AccessController.doPrivileged(Native
>> >>> >> >>Method)
>> >>> >> >> >> >>         at
>> >>>javax.security.auth.Subject.doAs(Subject.java:422)
>> >>> >> >> >> >>         at
>> >>> >> >> >> >>
>> >>> >> >> >>
>> >>> >> >>
>> >>> >>
>> >>>
>> >>>
>> 
>>>>>>>>>>>>>org.apache.hadoop.security.UserGroupInformation.doAs(UserGroup
>>>>>>>>>>>>>In
>> >>>>>>>>>>>for
>> >>> >>>>>>>>ma
>> >>> >> >>>>>>ti
>> >>> >> >> >>>>on
>> >>> >> >> >> >>.java:1671)
>> >>> >> >> >> >>         at
>> >>> >> >> >> >>
>> >>> >> >> >>
>> >>> >> >>
>> >>> >>
>> >>>
>> >>>
>> 
>>>>>>>>>>>>>org.apache.hadoop.ipc.Client$Connection.handleSaslConnectionFa
>>>>>>>>>>>>>il
>> >>>>>>>>>>>ure
>> >>> >>>>>>>>(C
>> >>> >> >>>>>>li
>> >>> >> >> >>>>en
>> >>> >> >> >> >>t.java:645)
>> >>> >> >> >> >>         at
>> >>> >> >> >> >>
>> >>> >> >>
>> >>> >>
>> >>>
>> >>>
>> 
>>>>>>>>>>>org.apache.hadoop.ipc.Client$Connection.setupIOstreams(Client.ja
>>>>>>>>>>>va
>> >>>>>>>>>:73
>> >>> >>>>>>3)
>> >>> >> >> >> >>         at
>> >>> >> >> >> >>
>> >>> >>
>> 
>>>>>>>org.apache.hadoop.ipc.Client$Connection.access$2800(Client.java:370)
>> >>> >> >> >> >>         at
>> >>> >> >> >>org.apache.hadoop.ipc.Client.getConnection(Client.java:1525)
>> >>> >> >> >> >>         at
>> >>>org.apache.hadoop.ipc.Client.call(Client.java:1442)
>> >>> >> >> >> >>         ... 12 more
>> >>> >> >> >> >> Caused by:
>> >>>org.apache.hadoop.security.AccessControlException:
>> >>> >> >>Client
>> >>> >> >> >> >> cannot authenticate via:[TOKEN]
>> >>> >> >> >> >>         at
>> >>> >> >> >> >>
>> >>> >> >> >>
>> >>> >> >>
>> >>> >>
>> >>>
>> >>>
>> 
>>>>>>>>>>>>>org.apache.hadoop.security.SaslRpcClient.selectSaslClient(Sasl
>>>>>>>>>>>>>Rp
>> >>>>>>>>>>>cCl
>> >>> >>>>>>>>ie
>> >>> >> >>>>>>nt
>> >>> >> >> >>>>.j
>> >>> >> >> >> >>ava:172)
>> >>> >> >> >> >>         at
>> >>> >> >> >> >>
>> >>> >> >> >>
>> >>> >> >>
>> >>> >>
>> >>>
>> >>>
>> 
>>>>>>>>>>>>>org.apache.hadoop.security.SaslRpcClient.saslConnect(SaslRpcCl
>>>>>>>>>>>>>ie
>> >>>>>>>>>>>nt.
>> >>> >>>>>>>>ja
>> >>> >> >>>>>>va
>> >>> >> >> >>>>:3
>> >>> >> >> >> >>96)
>> >>> >> >> >> >>         at
>> >>> >> >> >> >>
>> >>> >> >> >>
>> >>> >> >>
>> >>> >>
>> >>>
>> >>>
>> 
>>>>>>>>>>>>>org.apache.hadoop.ipc.Client$Connection.setupSaslConnection(Cl
>>>>>>>>>>>>>ie
>> >>>>>>>>>>>nt.
>> >>> >>>>>>>>ja
>> >>> >> >>>>>>va
>> >>> >> >> >>>>:5
>> >>> >> >> >> >>55)
>> >>> >> >> >> >>         at
>> >>> >> >> >> >>
>> >>> >>
>> 
>>>>>>>org.apache.hadoop.ipc.Client$Connection.access$1800(Client.java:370)
>> >>> >> >> >> >>         at
>> >>> >> >> >>
>> >>>>>org.apache.hadoop.ipc.Client$Connection$2.run(Client.java:725)
>> >>> >> >> >> >>         at
>> >>> >> >> >>
>> >>>>>org.apache.hadoop.ipc.Client$Connection$2.run(Client.java:721)
>> >>> >> >> >> >>         at
>> >>>java.security.AccessController.doPrivileged(Native
>> >>> >> >>Method)
>> >>> >> >> >> >>         at
>> >>>javax.security.auth.Subject.doAs(Subject.java:422)
>> >>> >> >> >> >>         at
>> >>> >> >> >> >>
>> >>> >> >> >>
>> >>> >> >>
>> >>> >>
>> >>>
>> >>>
>> 
>>>>>>>>>>>>>org.apache.hadoop.security.UserGroupInformation.doAs(UserGroup
>>>>>>>>>>>>>In
>> >>>>>>>>>>>for
>> >>> >>>>>>>>ma
>> >>> >> >>>>>>ti
>> >>> >> >> >>>>on
>> >>> >> >> >> >>.java:1671)
>> >>> >> >> >> >>         at
>> >>> >> >> >> >>
>> >>> >> >>
>> >>> >>
>> >>>
>> >>>
>> 
>>>>>>>>>>>org.apache.hadoop.ipc.Client$Connection.setupIOstreams(Client.ja
>>>>>>>>>>>va
>> >>>>>>>>>:72
>> >>> >>>>>>0)
>> >>> >> >> >> >>         ... 15 more
>> >>> >> >> >> >> 2016-07-25 19:06:48,089 [AMRM Heartbeater thread] INFO
>> >>> >> >> >> >>  client.ConfiguredRMFailoverProxyProvider - Failing
>>over to
>> >>> rm1
>> >>> >> >> >> >>
>> >>> >> >> >>
>> >>> >> >> >>
>> >>> >> >>
>> >>> >> >>
>> >>> >>
>> >>> >>
>> >>>
>> >>>
>> >>
>>
>>

Reply via email to