That is strange, since it is indeed not required to contain anything in slider-client.xml (except <configuration></configuration>) if HADOOP_CONF_DIR has everything that Slider needs. This probably gives an indication that there might be some issue with cluster configuration based on files solely under HADOOP_CONF_DIR to begin with.
Suggest you to upload all the config files to the jira to help debug this further. -Gour On 7/28/16, 4:27 PM, "Manoj Samel" <manojsamelt...@gmail.com> wrote: >Thanks Gour for prompt reply > >BTW - Creating a empty slider-client.xml (with just ><configuration></configuration>) does not works. The AM starts but fails >to >create any components and shows errors like > >2016-07-28 23:18:46,018 >[AmExecutor-006-SendThread(localhost.localdomain:2181)] WARN > zookeeper.ClientCnxn - Session 0x0 for server null, unexpected error, >closing socket connection and attempting reconnect >java.net.ConnectException: Connection refused > at sun.nio.ch.SocketChannelImpl.checkConnect(Native Method) > at >sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:717) > at >org.apache.zookeeper.ClientCnxnSocketNIO.doTransport(ClientCnxnSocketNIO.j >ava:361) > at >org.apache.zookeeper.ClientCnxn$SendThread.run(ClientCnxn.java:1081) > >Also, command "slider destroy <app>" fails with zookeeper errors ... > >I had to keep a minimal slider-client.xml. It does not have any RM info >etc. but does contain slider ZK related properties like >"slider.zookeeper.quorum", "hadoop.registry.zk.quorum", >"hadoop.registry.zk.root". I haven't yet distilled the absolute minimal >set >of properties required, but this should suffice for now. All RM / HDFS >properties will be read from HADOOP_CONF_DIR files. > >Let me know if this could cause any issues. > >On Thu, Jul 28, 2016 at 3:36 PM, Gour Saha <gs...@hortonworks.com> wrote: > >> No need to copy any files. Pointing HADOOP_CONF_DIR to /etc/hadoop/conf >>is >> good. >> >> -Gour >> >> On 7/28/16, 3:24 PM, "Manoj Samel" <manojsamelt...@gmail.com> wrote: >> >> >Follow up question regarding Gour's comment in earlier thread - >> > >> >Slider is installed on one of the hadoop nodes. SLIDER_HOME/conf >>directory >> >(say /data/slider/conf) is different than HADOOP_CONF_DIR >> >(/etc/hadoop/conf). Is it required/recommended that files in >> >HADOOP_CONF_DIR be copied to SLIDER_HOME/conf and slider-env.sh script >> >sets >> >HADOOP_CONF_DIR to /data/slider/conf ? >> > >> >Or can the slider-env.sh set HADOOP_CONF_DIR to /etc/hadoop/conf , >>without >> >copying the files ? >> > >> >Using slider .80 for now, but would like to know recommendation for >>this >> >and future versions as well. >> > >> >Thanks in advance, >> > >> >Manoj >> > >> >On Tue, Jul 26, 2016 at 3:27 PM, Manoj Samel <manojsamelt...@gmail.com> >> >wrote: >> > >> >> Filed https://issues.apache.org/jira/browse/SLIDER-1158 with logs and >> my >> >> analysis of logs. >> >> >> >> On Tue, Jul 26, 2016 at 10:36 AM, Gour Saha <gs...@hortonworks.com> >> >>wrote: >> >> >> >>> Please file a JIRA and upload the logs to it. >> >>> >> >>> On 7/26/16, 10:21 AM, "Manoj Samel" <manojsamelt...@gmail.com> >>wrote: >> >>> >> >>> >Hi Gour, >> >>> > >> >>> >Can you please reach me using your own email-id? I will then send >> >>>logs to >> >>> >you, along with my analysis - I don't want to send logs on public >>list >> >>> > >> >>> >Thanks, >> >>> > >> >>> >On Mon, Jul 25, 2016 at 5:39 PM, Gour Saha <gs...@hortonworks.com> >> >>> wrote: >> >>> > >> >>> >> Ok, so this node is not a gateway. It is part of the cluster, >>which >> >>> >>means >> >>> >> you don¹t need slider-client.xml at all. Just have >>HADOOP_CONF_DIR >> >>> >> pointing to /etc/hadoop/conf in slider-env.sh and that should be >>it. >> >>> >> >> >>> >> So the above simplifies your config setup. It will not solve >>either >> >>>of >> >>> >>the >> >>> >> 2 problems you are facing. >> >>> >> >> >>> >> Now coming to the 2 issues you are facing, you have to provide >> >>> >>additional >> >>> >> logs for us to understand better. Let¹s start with - >> >>> >> 1. RM logs (specifically between the time when rm1->rm2 failover >>is >> >>> >> simulated) >> >>> >> 2. Slider App logs >> >>> >> >> >>> >> -Gour >> >>> >> >> >>> >> On 7/25/16, 5:16 PM, "Manoj Samel" <manojsamelt...@gmail.com> >> wrote: >> >>> >> >> >>> >> > 1. Not clear about your question on "gateway" node. The node >> >>> running >> >>> >> > slider is part of the hadoop cluster and there are other >> >>>services >> >>> >>like >> >>> >> > Oozie that run on this node that utilizes hdfs and yarn. So >>if >> >>>your >> >>> >> > question is whether the node is otherwise working for HDFS >>and >> >>>Yarn >> >>> >> > configuration, it is working >> >>> >> > 2. I copied all files from HADOOP_CONF_DIR (say >> >>>/etc/hadoop/conf) >> >>> to >> >>> >> >the >> >>> >> > directory containing slider-client.xml (say >>/data/latest/conf) >> >>> >> > 3. In earlier email, I had done a mistake where slider-env.sh >> >>>file >> >>> >> >HADOOP_CONF_DIR >> >>> >> > was pointing to original directory /etc/hadoop/conf. I edited >> >>>it to >> >>> >> > point to same directory containing slider-client.xml & >> >>> slider-env.sh >> >>> >> >i.e. >> >>> >> > /data/latest/conf >> >>> >> > 4. I emptied slider-client.xml. It just had the >> >>> >> ><configuration></configuration>. >> >>> >> > The creation of spas worked but the Slider AM still shows the >> >>>same >> >>> >> >issue. >> >>> >> > i.e. when RM1 goes from active to standby, slider AM goes >>from >> >>> >>RUNNING >> >>> >> >to >> >>> >> > ACCPTED state with same error about TOKEN. Also NOTE that >>when >> >>> >> > slider-client.xml is empty, the "slider destroy xxx" command >> >>>still >> >>> >> >fails >> >>> >> > with Zookeeper connection errors. >> >>> >> > 5. I then added same parameters (as my last email - except >> >>> >> > HADOOP_CONF_DIR) to slider-client.xml and ran. This time >> >>> >>slider-env.sh >> >>> >> > has HADOOP_CONF_DIR pointing to /data/latest/conf and >> >>> >>slider-client.xml >> >>> >> > does not have HADOOP_CONF_DIR. The same issue exists (but >> >>>"slider >> >>> >> > destroy" does not fails) >> >>> >> > 6. Could you explain what do you expect to pick up from >>Hadoop >> >>> >> > configurations that will help you in RM Token ? If slider has >> >>>token >> >>> >> >from >> >>> >> > RM1, and it switches to RM2, not clear what slider does to >>get >> >>> >> >delegation >> >>> >> > token for RM2 communication ? >> >>> >> > 7. It is worth repeating again that issue happens only when >>RM1 >> >>>was >> >>> >> > active when slider app was created and then RM1 becomes >> >>>standby. If >> >>> >> >RM2 was >> >>> >> > active when slider app was created, then slider AM keeps >>running >> >>> for >> >>> >> >any >> >>> >> > number of switches between RM2 and RM1 back and forth ... >> >>> >> > >> >>> >> > >> >>> >> >On Mon, Jul 25, 2016 at 4:21 PM, Gour Saha >><gs...@hortonworks.com> >> >>> >>wrote: >> >>> >> > >> >>> >> >> The node you are running slider from, is that a gateway node? >> >>>Sorry >> >>> >>for >> >>> >> >> not being explicit. I meant copy everything under >> >>>/etc/hadoop/conf >> >>> >>from >> >>> >> >> your cluster into some temp directory (say /tmp/hadoop_conf) >>in >> >>>your >> >>> >> >> gateway node or local or whichever node you are running slider >> >>>from. >> >>> >> >>Then >> >>> >> >> set HADOOP_CONF_DIR to /tmp/hadoop_conf and clear everything >>out >> >>> from >> >>> >> >> slider-client.xml. >> >>> >> >> >> >>> >> >> On 7/25/16, 4:12 PM, "Manoj Samel" <manojsamelt...@gmail.com> >> >>> wrote: >> >>> >> >> >> >>> >> >> >Hi Gour, >> >>> >> >> > >> >>> >> >> >Thanks for your prompt reply. >> >>> >> >> > >> >>> >> >> >FYI, issue happens when I create slider app when rm1 is >>active >> >>>and >> >>> >>when >> >>> >> >> >rm1 >> >>> >> >> >fails over to rm2. As soon as rm2 becomes active; the slider >>AM >> >>> goes >> >>> >> >>from >> >>> >> >> >RUNNING to ACCEPTED state with above error. >> >>> >> >> > >> >>> >> >> >For your suggestion, I did following >> >>> >> >> > >> >>> >> >> >1) Copied core-site, hdfs-site, yarn-site, and mapred-site >>from >> >>> >> >> >HADOOP_CONF_DIR >> >>> >> >> >to slider conf directory. >> >>> >> >> >2) Our slider-env.sh already had HADOOP_CONF_DIR set >> >>> >> >> >3) I removed all properties from slider-client.xml EXCEPT >> >>>following >> >>> >> >> > >> >>> >> >> > - HADOOP_CONF_DIR >> >>> >> >> > - slider.yarn.queue >> >>> >> >> > - slider.zookeeper.quorum >> >>> >> >> > - hadoop.registry.zk.quorum >> >>> >> >> > - hadoop.registry.zk.root >> >>> >> >> > - hadoop.security.authorization >> >>> >> >> > - hadoop.security.authentication >> >>> >> >> > >> >>> >> >> >Then I made rm1 active, installed and created slider app and >> >>> >>restarted >> >>> >> >>rm1 >> >>> >> >> >(to make rm2) active. The slider-am again went from RUNNING >>to >> >>> >>ACCEPTED >> >>> >> >> >state. >> >>> >> >> > >> >>> >> >> >Let me know if you want me to try further changes. >> >>> >> >> > >> >>> >> >> >If I make the slider-client.xml completely empty per your >> >>> >>suggestion, >> >>> >> >>only >> >>> >> >> >slider AM comes up but it >> >>> >> >> >fails to start components. The AM log shows errors trying to >> >>> >>connect to >> >>> >> >> >zookeeper like below. >> >>> >> >> >2016-07-25 23:07:41,532 >> >>> >> >> >[AmExecutor-006-SendThread(localhost.localdomain:2181)] WARN >> >>> >> >> >zookeeper.ClientCnxn - Session 0x0 for server null, >>unexpected >> >>> >>error, >> >>> >> >> >closing socket connection and attempting reconnect >> >>> >> >> >java.net.ConnectException: Connection refused >> >>> >> >> > >> >>> >> >> >Hence I kept minimal info in slider-client.xml >> >>> >> >> > >> >>> >> >> >FYI This is slider version 0.80 >> >>> >> >> > >> >>> >> >> >Thanks, >> >>> >> >> > >> >>> >> >> >Manoj >> >>> >> >> > >> >>> >> >> >On Mon, Jul 25, 2016 at 2:54 PM, Gour Saha >> >>><gs...@hortonworks.com> >> >>> >> >>wrote: >> >>> >> >> > >> >>> >> >> >> If possible, can you copy the entire content of the >>directory >> >>> >> >> >> /etc/hadoop/conf and then set HADOOP_CONF_DIR in >> >>>slider-env.sh to >> >>> >>it. >> >>> >> >> >>Keep >> >>> >> >> >> slider-client.xml empty. >> >>> >> >> >> >> >>> >> >> >> Now when you do the same rm1->rm2 and then the reverse >> >>>failovers, >> >>> >>do >> >>> >> >>you >> >>> >> >> >> see the same behaviors? >> >>> >> >> >> >> >>> >> >> >> -Gour >> >>> >> >> >> >> >>> >> >> >> On 7/25/16, 2:28 PM, "Manoj Samel" >><manojsamelt...@gmail.com> >> >>> >>wrote: >> >>> >> >> >> >> >>> >> >> >> >Another observation (whatever it is worth) >> >>> >> >> >> > >> >>> >> >> >> >If slider app is created and started when rm2 was active, >> >>>then >> >>> it >> >>> >> >> >>seems to >> >>> >> >> >> >survive switches between rm2 and rm1 (and back). I.e >> >>> >> >> >> > >> >>> >> >> >> >* rm2 is active >> >>> >> >> >> >* create and start slider application >> >>> >> >> >> >* fail over to rm1. Now the Slider AM keeps running >> >>> >> >> >> >* fail over to rm2 again. Slider AM still keeps running >> >>> >> >> >> > >> >>> >> >> >> >So, it seems if it starts with rm1 active, then the AM >>goes >> >>>to >> >>> >> >> >>"ACCEPTED" >> >>> >> >> >> >state when RM fails to rm2. If it starts with rm2 active, >> >>>then >> >>> it >> >>> >> >>runs >> >>> >> >> >> >fine >> >>> >> >> >> >with any switches between rm1 and rm2. >> >>> >> >> >> > >> >>> >> >> >> >Any feedback ? >> >>> >> >> >> > >> >>> >> >> >> >Thanks, >> >>> >> >> >> > >> >>> >> >> >> >Manoj >> >>> >> >> >> > >> >>> >> >> >> >On Mon, Jul 25, 2016 at 12:25 PM, Manoj Samel >> >>> >> >> >><manojsamelt...@gmail.com> >> >>> >> >> >> >wrote: >> >>> >> >> >> > >> >>> >> >> >> >> Setup >> >>> >> >> >> >> >> >>> >> >> >> >> - Hadoop 2.6 with RM HA, Kerberos enabled >> >>> >> >> >> >> - Slider 0.80 >> >>> >> >> >> >> - In my slider-client.xml, I have added all RM HA >> >>>properties, >> >>> >> >> >>including >> >>> >> >> >> >> the ones mentioned in >> >>> >> >>http://markmail.org/message/wnhpp2zn6ixo65e3. >> >>> >> >> >> >> >> >>> >> >> >> >> Following is the issue >> >>> >> >> >> >> >> >>> >> >> >> >> * rm1 is active, rm2 is standby >> >>> >> >> >> >> * deploy and start slider application, it runs fine >> >>> >> >> >> >> * restart rm1, rm2 is now active. >> >>> >> >> >> >> * The slider-am now goes from running into "ACCEPTED" >> >>>mode. It >> >>> >> >>stays >> >>> >> >> >> >>there >> >>> >> >> >> >> till rm1 is made active again. >> >>> >> >> >> >> >> >>> >> >> >> >> In the slider-am log, it tries to connect to RM2 and >> >>> connection >> >>> >> >>fails >> >>> >> >> >> >>due >> >>> >> >> >> >> to org.apache.hadoop.security.AccessControlException: >> >>>Client >> >>> >> >>cannot >> >>> >> >> >> >> authenticate via:[TOKEN]. See detailed log below >> >>> >> >> >> >> >> >>> >> >> >> >> It seems it has some token (delegation token?) for RM1 >>but >> >>> >>tries >> >>> >> >>to >> >>> >> >> >>use >> >>> >> >> >> >> same(?) for RM2 and fails. Am I missing some >>configuration >> >>>??? >> >>> >> >> >> >> >> >>> >> >> >> >> Thanks, >> >>> >> >> >> >> >> >>> >> >> >> >> >> >>> >> >> >> >> >> >>> >> >> >> >> 2016-07-25 19:06:48,088 [AMRM Heartbeater thread] INFO >> >>> >> >> >> >> client.ConfiguredRMFailoverProxyProvider - Failing >>over to >> >>> rm2 >> >>> >> >> >> >> 2016-07-25 19:06:48,088 [AMRM Heartbeater thread] WARN >> >>> >> >> >> >> security.UserGroupInformation - >>PriviledgedActionException >> >>> >> >> >>as:abc@XYZ >> >>> >> >> >> >> (auth:KERBEROS) >> >>> >> >> >>cause:org.apache.hadoop.security.AccessControlException: >> >>> >> >> >> >> Client cannot authenticate via:[TOKEN] >> >>> >> >> >> >> 2016-07-25 19:06:48,088 [AMRM Heartbeater thread] WARN >> >>> >> >>ipc.Client - >> >>> >> >> >> >> Exception encountered while connecting to the server : >> >>> >> >> >> >> org.apache.hadoop.security.AccessControlException: >>Client >> >>> >>cannot >> >>> >> >> >> >> authenticate via:[TOKEN] >> >>> >> >> >> >> 2016-07-25 19:06:48,088 [AMRM Heartbeater thread] WARN >> >>> >> >> >> >> security.UserGroupInformation - >>PriviledgedActionException >> >>> >> >> >>as:abc@XYZ >> >>> >> >> >> >> (auth:KERBEROS) cause:java.io.IOException: >> >>> >> >> >> >> org.apache.hadoop.security.AccessControlException: >>Client >> >>> >>cannot >> >>> >> >> >> >> authenticate via:[TOKEN] >> >>> >> >> >> >> 2016-07-25 19:06:48,089 [AMRM Heartbeater thread] INFO >> >>> >> >> >> >> retry.RetryInvocationHandler - Exception while invoking >> >>> >>allocate >> >>> >> >>of >> >>> >> >> >> >>class >> >>> >> >> >> >> ApplicationMasterProtocolPBClientImpl over rm2 after 287 >> >>>fail >> >>> >>over >> >>> >> >> >> >> attempts. Trying to fail over immediately. >> >>> >> >> >> >> java.io.IOException: Failed on local exception: >> >>> >> >>java.io.IOException: >> >>> >> >> >> >> org.apache.hadoop.security.AccessControlException: >>Client >> >>> >>cannot >> >>> >> >> >> >> authenticate via:[TOKEN]; Host Details : local host is: >> >>> >>"<SliderAM >> >>> >> >> >> >> HOST>/<slider AM Host IP>"; destination host is: "<RM2 >> >>> >> >>HOST>":23130; >> >>> >> >> >> >> at >> >>> >> >> >> >> >>>>>org.apache.hadoop.net.NetUtils.wrapException(NetUtils.java:772) >> >>> >> >> >> >> at >> >>>org.apache.hadoop.ipc.Client.call(Client.java:1476) >> >>> >> >> >> >> at >> >>>org.apache.hadoop.ipc.Client.call(Client.java:1403) >> >>> >> >> >> >> at >> >>> >> >> >> >> >> >>> >> >> >> >> >>> >> >> >> >>> >> >> >>> >> >>> >> >>>>>>>>>>>>>org.apache.hadoop.ipc.ProtobufRpcEngine$Invoker.invoke(Protobu >>>>>>>>>>>>>fR >> >>>>>>>>>>>pcE >> >>> >>>>>>>>ng >> >>> >> >>>>>>in >> >>> >> >> >>>>e. >> >>> >> >> >> >>java:230) >> >>> >> >> >> >> at com.sun.proxy.$Proxy23.allocate(Unknown >>Source) >> >>> >> >> >> >> at >> >>> >> >> >> >> >> >>> >> >> >> >> >>> >> >> >> >>> >> >> >>> >> >>> >> >>>>>>>>>>>>>org.apache.hadoop.yarn.api.impl.pb.client.ApplicationMasterPro >>>>>>>>>>>>>to >> >>>>>>>>>>>col >> >>> >>>>>>>>PB >> >>> >> >>>>>>Cl >> >>> >> >> >>>>ie >> >>> >> >> >> >> >>>>>ntImpl.allocate(ApplicationMasterProtocolPBClientImpl.java:77) >> >>> >> >> >> >> at >> >>> sun.reflect.GeneratedMethodAccessor10.invoke(Unknown >> >>> >> >> >>Source) >> >>> >> >> >> >> at >> >>> >> >> >> >> >> >>> >> >> >> >> >>> >> >> >> >>> >> >> >>> >> >>> >> >>>>>>>>>>>>>sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMeth >>>>>>>>>>>>>od >> >>>>>>>>>>>Acc >> >>> >>>>>>>>es >> >>> >> >>>>>>so >> >>> >> >> >>>>rI >> >>> >> >> >> >>mpl.java:43) >> >>> >> >> >> >> at >>java.lang.reflect.Method.invoke(Method.java:497) >> >>> >> >> >> >> at >> >>> >> >> >> >> >> >>> >> >> >> >> >>> >> >> >> >>> >> >> >>> >> >>> >> >>>>>>>>>>>>>org.apache.hadoop.io.retry.RetryInvocationHandler.invokeMethod >>>>>>>>>>>>>(R >> >>>>>>>>>>>etr >> >>> >>>>>>>>yI >> >>> >> >>>>>>nv >> >>> >> >> >>>>oc >> >>> >> >> >> >>ationHandler.java:252) >> >>> >> >> >> >> at >> >>> >> >> >> >> >> >>> >> >> >> >> >>> >> >> >> >>> >> >> >>> >> >>> >> >>>>>>>>>>>>>org.apache.hadoop.io.retry.RetryInvocationHandler.invoke(Retry >>>>>>>>>>>>>In >> >>>>>>>>>>>voc >> >>> >>>>>>>>at >> >>> >> >>>>>>io >> >>> >> >> >>>>nH >> >>> >> >> >> >>andler.java:104) >> >>> >> >> >> >> at com.sun.proxy.$Proxy24.allocate(Unknown >>Source) >> >>> >> >> >> >> at >> >>> >> >> >> >> >> >>> >> >> >> >> >>> >> >> >> >>> >> >> >>> >> >>> >> >>>>>>>>>>>>>org.apache.hadoop.yarn.client.api.impl.AMRMClientImpl.allocate >>>>>>>>>>>>>(A >> >>>>>>>>>>>MRM >> >>> >>>>>>>>Cl >> >>> >> >>>>>>ie >> >>> >> >> >>>>nt >> >>> >> >> >> >>Impl.java:278) >> >>> >> >> >> >> at >> >>> >> >> >> >> >> >>> >> >> >> >> >>> >> >> >> >>> >> >> >>> >> >>> >> >>>>>>>>>>>>>org.apache.hadoop.yarn.client.api.async.impl.AMRMClientAsyncIm >>>>>>>>>>>>>pl >> >>>>>>>>>>>$He >> >>> >>>>>>>>ar >> >>> >> >>>>>>tb >> >>> >> >> >>>>ea >> >>> >> >> >> >>tThread.run(AMRMClientAsyncImpl.java:224) >> >>> >> >> >> >> Caused by: java.io.IOException: >> >>> >> >> >> >> org.apache.hadoop.security.AccessControlException: >>Client >> >>> >>cannot >> >>> >> >> >> >> authenticate via:[TOKEN] >> >>> >> >> >> >> at >> >>> >> >> >> >> >>>>>org.apache.hadoop.ipc.Client$Connection$1.run(Client.java:682) >> >>> >> >> >> >> at >> >>>java.security.AccessController.doPrivileged(Native >> >>> >> >>Method) >> >>> >> >> >> >> at >> >>>javax.security.auth.Subject.doAs(Subject.java:422) >> >>> >> >> >> >> at >> >>> >> >> >> >> >> >>> >> >> >> >> >>> >> >> >> >>> >> >> >>> >> >>> >> >>>>>>>>>>>>>org.apache.hadoop.security.UserGroupInformation.doAs(UserGroup >>>>>>>>>>>>>In >> >>>>>>>>>>>for >> >>> >>>>>>>>ma >> >>> >> >>>>>>ti >> >>> >> >> >>>>on >> >>> >> >> >> >>.java:1671) >> >>> >> >> >> >> at >> >>> >> >> >> >> >> >>> >> >> >> >> >>> >> >> >> >>> >> >> >>> >> >>> >> >>>>>>>>>>>>>org.apache.hadoop.ipc.Client$Connection.handleSaslConnectionFa >>>>>>>>>>>>>il >> >>>>>>>>>>>ure >> >>> >>>>>>>>(C >> >>> >> >>>>>>li >> >>> >> >> >>>>en >> >>> >> >> >> >>t.java:645) >> >>> >> >> >> >> at >> >>> >> >> >> >> >> >>> >> >> >> >>> >> >> >>> >> >>> >> >>>>>>>>>>>org.apache.hadoop.ipc.Client$Connection.setupIOstreams(Client.ja >>>>>>>>>>>va >> >>>>>>>>>:73 >> >>> >>>>>>3) >> >>> >> >> >> >> at >> >>> >> >> >> >> >> >>> >> >> >>>>>>>org.apache.hadoop.ipc.Client$Connection.access$2800(Client.java:370) >> >>> >> >> >> >> at >> >>> >> >> >>org.apache.hadoop.ipc.Client.getConnection(Client.java:1525) >> >>> >> >> >> >> at >> >>>org.apache.hadoop.ipc.Client.call(Client.java:1442) >> >>> >> >> >> >> ... 12 more >> >>> >> >> >> >> Caused by: >> >>>org.apache.hadoop.security.AccessControlException: >> >>> >> >>Client >> >>> >> >> >> >> cannot authenticate via:[TOKEN] >> >>> >> >> >> >> at >> >>> >> >> >> >> >> >>> >> >> >> >> >>> >> >> >> >>> >> >> >>> >> >>> >> >>>>>>>>>>>>>org.apache.hadoop.security.SaslRpcClient.selectSaslClient(Sasl >>>>>>>>>>>>>Rp >> >>>>>>>>>>>cCl >> >>> >>>>>>>>ie >> >>> >> >>>>>>nt >> >>> >> >> >>>>.j >> >>> >> >> >> >>ava:172) >> >>> >> >> >> >> at >> >>> >> >> >> >> >> >>> >> >> >> >> >>> >> >> >> >>> >> >> >>> >> >>> >> >>>>>>>>>>>>>org.apache.hadoop.security.SaslRpcClient.saslConnect(SaslRpcCl >>>>>>>>>>>>>ie >> >>>>>>>>>>>nt. >> >>> >>>>>>>>ja >> >>> >> >>>>>>va >> >>> >> >> >>>>:3 >> >>> >> >> >> >>96) >> >>> >> >> >> >> at >> >>> >> >> >> >> >> >>> >> >> >> >> >>> >> >> >> >>> >> >> >>> >> >>> >> >>>>>>>>>>>>>org.apache.hadoop.ipc.Client$Connection.setupSaslConnection(Cl >>>>>>>>>>>>>ie >> >>>>>>>>>>>nt. >> >>> >>>>>>>>ja >> >>> >> >>>>>>va >> >>> >> >> >>>>:5 >> >>> >> >> >> >>55) >> >>> >> >> >> >> at >> >>> >> >> >> >> >> >>> >> >> >>>>>>>org.apache.hadoop.ipc.Client$Connection.access$1800(Client.java:370) >> >>> >> >> >> >> at >> >>> >> >> >> >> >>>>>org.apache.hadoop.ipc.Client$Connection$2.run(Client.java:725) >> >>> >> >> >> >> at >> >>> >> >> >> >> >>>>>org.apache.hadoop.ipc.Client$Connection$2.run(Client.java:721) >> >>> >> >> >> >> at >> >>>java.security.AccessController.doPrivileged(Native >> >>> >> >>Method) >> >>> >> >> >> >> at >> >>>javax.security.auth.Subject.doAs(Subject.java:422) >> >>> >> >> >> >> at >> >>> >> >> >> >> >> >>> >> >> >> >> >>> >> >> >> >>> >> >> >>> >> >>> >> >>>>>>>>>>>>>org.apache.hadoop.security.UserGroupInformation.doAs(UserGroup >>>>>>>>>>>>>In >> >>>>>>>>>>>for >> >>> >>>>>>>>ma >> >>> >> >>>>>>ti >> >>> >> >> >>>>on >> >>> >> >> >> >>.java:1671) >> >>> >> >> >> >> at >> >>> >> >> >> >> >> >>> >> >> >> >>> >> >> >>> >> >>> >> >>>>>>>>>>>org.apache.hadoop.ipc.Client$Connection.setupIOstreams(Client.ja >>>>>>>>>>>va >> >>>>>>>>>:72 >> >>> >>>>>>0) >> >>> >> >> >> >> ... 15 more >> >>> >> >> >> >> 2016-07-25 19:06:48,089 [AMRM Heartbeater thread] INFO >> >>> >> >> >> >> client.ConfiguredRMFailoverProxyProvider - Failing >>over to >> >>> rm1 >> >>> >> >> >> >> >> >>> >> >> >> >> >>> >> >> >> >> >>> >> >> >> >>> >> >> >> >>> >> >> >>> >> >> >>> >> >>> >> >> >> >>